++ LOG_DIR=/var/log/contrail ++ export CONTAINER_LOG_DIR=/var/log/contrail/config-database-cassandra ++ CONTAINER_LOG_DIR=/var/log/contrail/config-database-cassandra ++ mkdir -p /var/log/contrail/config-database-cassandra ++ log_file=/var/log/contrail/config-database-cassandra/console.log ++ touch /var/log/contrail/config-database-cassandra/console.log ++ chmod 600 /var/log/contrail/config-database-cassandra/console.log ++ exec +++ tee -a /var/log/contrail/config-database-cassandra/console.log +++ date ++ echo 'INFO: =================== Thu Oct 16 05:22:31 UTC 2025 ===================' INFO: =================== Thu Oct 16 05:22:31 UTC 2025 =================== ++ LOG_LOCAL=1 ++ source /functions.sh ++ source /contrail-functions.sh +++ get_default_ip ++++ get_default_nic ++++ get_gateway_nic_for_ip 1 ++++ command -v ip ++++ local ip=1 +++++ ip route get 1 +++++ grep -o 'dev.*' +++++ awk '{print $2}' ++++ local iface=ens3 ++++ [[ ens3 == \l\o ]] ++++ echo ens3 +++ local nic=ens3 +++ get_ip_for_nic ens3 +++ local nic=ens3 +++ get_cidr_for_nic ens3 +++ command -v ip +++ local nic=ens3 +++ ip addr show dev ens3 +++ grep 'inet ' +++ awk '{print $2}' +++ head -n 1 +++ cut -d / -f 1 ++ DEFAULT_LOCAL_IP=10.0.0.242 ++ ENCAP_PRIORITY=MPLSoUDP,MPLSoGRE,VXLAN ++ VXLAN_VN_ID_MODE=automatic ++ DPDK_UIO_DRIVER=uio_pci_generic ++ CPU_CORE_MASK=0x01 ++ SERVICE_CORE_MASK= ++ DPDK_CTRL_THREAD_MASK= ++ HUGE_PAGES= ++ HUGE_PAGES_DIR=/dev/hugepages ++ HUGE_PAGES_1GB=0 ++ HUGE_PAGES_2MB=256 ++ HUGE_PAGES_1GB_DIR= ++ HUGE_PAGES_2MB_DIR= ++ [[ 0 != 0 ]] ++ [[ 0 != 256 ]] ++ [[ -z '' ]] +++ awk '/pagesize=2M/{print($3)}' +++ tail -n 1 +++ mount -t hugetlbfs ++ HUGE_PAGES_2MB_DIR= ++ DPDK_MEM_PER_SOCKET=1024 ++ DPDK_COMMAND_ADDITIONAL_ARGS= ++ NIC_OFFLOAD_ENABLE=False ++ DPDK_ENABLE_VLAN_FWRD=False ++ DIST_SNAT_PROTO_PORT_LIST= ++ CLOUD_ORCHESTRATOR=openstack ++ CLOUD_ADMIN_ROLE=admin ++ AAA_MODE=rbac ++ AUTH_MODE=keystone ++ AUTH_PARAMS= ++ SSL_ENABLE=false ++ SSL_INSECURE=True ++ SERVER_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ SERVER_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ SERVER_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ SERVER_CA_KEYFILE=/etc/contrail/ssl/private/ca-key.pem ++ SELFSIGNED_CERTS_WITH_IPS=True ++ CONTROLLER_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ANALYTICS_ALARM_ENABLE=True ++ ANALYTICS_SNMP_ENABLE=True ++ ANALYTICSDB_ENABLE=True ++ ANALYTICS_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ANALYTICSDB_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ANALYTICS_SNMP_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ANALYTICS_API_PORT=8081 ++ ANALYTICS_API_INTROSPECT_PORT=8090 ++ ANALYTICSDB_PORT=9160 ++ ANALYTICSDB_CQL_PORT=9042 ++ TOPOLOGY_INTROSPECT_PORT=5921 ++ QUERYENGINE_INTROSPECT_PORT=8091 +++ get_server_list ANALYTICS ':8081 ' +++ local server_typ=ANALYTICS_NODES +++ local 'port_with_delim=:8081 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:8081 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:8081 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:8081 ' +++ '[' -n '10.0.0.38:8081 10.0.0.241:8081 10.0.0.242:8081 ' ']' +++ echo '10.0.0.38:8081 10.0.0.241:8081 10.0.0.242:8081' ++ ANALYTICS_SERVERS='10.0.0.38:8081 10.0.0.241:8081 10.0.0.242:8081' +++ get_server_list ANALYTICSDB ':9042 ' +++ local server_typ=ANALYTICSDB_NODES +++ local 'port_with_delim=:9042 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:9042 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:9042 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:9042 ' +++ '[' -n '10.0.0.38:9042 10.0.0.241:9042 10.0.0.242:9042 ' ']' +++ echo '10.0.0.38:9042 10.0.0.241:9042 10.0.0.242:9042' ++ ANALYTICSDB_CQL_SERVERS='10.0.0.38:9042 10.0.0.241:9042 10.0.0.242:9042' ++ ANALYTICS_API_VIP= ++ ANALYTICS_ALARM_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ALARMGEN_INTROSPECT_PORT=5995 ++ BGP_PORT=179 ++ BGP_AUTO_MESH=true ++ BGP_ASN=64512 ++ ENABLE_4BYTE_AS=false ++ APPLY_DEFAULTS=true ++ COLLECTOR_PORT=8086 ++ COLLECTOR_INTROSPECT_PORT=8089 ++ COLLECTOR_SYSLOG_PORT=514 ++ COLLECTOR_SFLOW_PORT=6343 ++ COLLECTOR_IPFIX_PORT=4739 ++ COLLECTOR_PROTOBUF_PORT=3333 ++ COLLECTOR_STRUCTURED_SYSLOG_PORT=3514 ++ SNMPCOLLECTOR_INTROSPECT_PORT=5920 +++ get_server_list ANALYTICS ':8086 ' +++ local server_typ=ANALYTICS_NODES +++ local 'port_with_delim=:8086 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:8086 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:8086 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:8086 ' +++ '[' -n '10.0.0.38:8086 10.0.0.241:8086 10.0.0.242:8086 ' ']' +++ echo '10.0.0.38:8086 10.0.0.241:8086 10.0.0.242:8086' ++ COLLECTOR_SERVERS='10.0.0.38:8086 10.0.0.241:8086 10.0.0.242:8086' ++ CASSANDRA_PORT=9161 ++ CASSANDRA_CQL_PORT=9041 ++ CASSANDRA_SSL_STORAGE_PORT=7013 ++ CASSANDRA_STORAGE_PORT=7012 ++ CASSANDRA_JMX_LOCAL_PORT=7201 ++ CONFIGDB_CASSANDRA_DRIVER=cql ++ CONFIG_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ CONFIGDB_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ CONFIG_API_PORT=8082 ++ CONFIG_API_INTROSPECT_PORT=8084 ++ CONFIG_API_ADMIN_PORT=8095 ++ CONFIGDB_PORT=9161 ++ CONFIGDB_CQL_PORT=9041 +++ get_server_list CONFIG ':8082 ' +++ local server_typ=CONFIG_NODES +++ local 'port_with_delim=:8082 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:8082 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:8082 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:8082 ' +++ '[' -n '10.0.0.38:8082 10.0.0.241:8082 10.0.0.242:8082 ' ']' +++ echo '10.0.0.38:8082 10.0.0.241:8082 10.0.0.242:8082' ++ CONFIG_SERVERS='10.0.0.38:8082 10.0.0.241:8082 10.0.0.242:8082' +++ get_server_list CONFIGDB ':9161 ' +++ local server_typ=CONFIGDB_NODES +++ local 'port_with_delim=:9161 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:9161 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:9161 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:9161 ' +++ '[' -n '10.0.0.38:9161 10.0.0.241:9161 10.0.0.242:9161 ' ']' +++ echo '10.0.0.38:9161 10.0.0.241:9161 10.0.0.242:9161' ++ CONFIGDB_SERVERS='10.0.0.38:9161 10.0.0.241:9161 10.0.0.242:9161' +++ get_server_list CONFIGDB ':9041 ' +++ local server_typ=CONFIGDB_NODES +++ local 'port_with_delim=:9041 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:9041 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:9041 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:9041 ' +++ '[' -n '10.0.0.38:9041 10.0.0.241:9041 10.0.0.242:9041 ' ']' +++ echo '10.0.0.38:9041 10.0.0.241:9041 10.0.0.242:9041' ++ CONFIGDB_CQL_SERVERS='10.0.0.38:9041 10.0.0.241:9041 10.0.0.242:9041' ++ CONFIG_API_VIP= ++ CONFIG_API_SSL_ENABLE=false ++ CONFIG_API_SERVER_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ CONFIG_API_SERVER_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ CONFIG_API_SERVER_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ CONFIG_API_WORKER_COUNT=1 ++ CONFIG_API_MAX_REQUESTS=1024 ++ ANALYTICS_API_SSL_ENABLE=false ++ ANALYTICS_API_SSL_INSECURE=True ++ ANALYTICS_API_SERVER_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ ANALYTICS_API_SERVER_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ ANALYTICS_API_SERVER_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ CASSANDRA_SSL_ENABLE=false ++ CASSANDRA_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ CASSANDRA_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ CASSANDRA_SSL_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ CASSANDRA_SSL_KEYSTORE_PASSWORD=astrophytum ++ CASSANDRA_SSL_TRUSTSTORE_PASSWORD=ornatum ++ CASSANDRA_SSL_PROTOCOL=TLS ++ CASSANDRA_SSL_ALGORITHM=SunX509 ++ CASSANDRA_SSL_CIPHER_SUITES='[TLS_RSA_WITH_AES_128_CBC_SHA,TLS_RSA_WITH_AES_256_CBC_SHA,TLS_DHE_RSA_WITH_AES_128_CBC_SHA,TLS_DHE_RSA_WITH_AES_256_CBC_SHA,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA]' ++ CASSANDRA_CONFIG_MEMTABLE_FLUSH_WRITER=4 ++ CASSANDRA_CONFIG_CONCURRECT_COMPACTORS=4 ++ CASSANDRA_CONFIG_COMPACTION_THROUGHPUT_MB_PER_SEC=256 ++ CASSANDRA_CONFIG_CONCURRECT_READS=64 ++ CASSANDRA_CONFIG_CONCURRECT_WRITES=64 ++ CASSANDRA_CONFIG_MEMTABLE_ALLOCATION_TYPE=offheap_objects ++ CASSANDRA_REAPER_ENABLED=true ++ CASSANDRA_REAPER_JMX_KEY=reaperJmxKey ++ CASSANDRA_REAPER_JMX_AUTH_USERNAME=reaperUser ++ CASSANDRA_REAPER_JMX_AUTH_PASSWORD=reaperPass ++ CASSANDRA_REAPER_APP_PORT=8071 ++ CASSANDRA_REAPER_ADM_PORT=8072 ++ CONTROL_NODES=10.20.0.14,10.20.0.18,10.20.0.252 ++ CONTROL_INTROSPECT_PORT=8083 ++ DNS_NODES=10.20.0.14,10.20.0.18,10.20.0.252 ++ DNS_SERVER_PORT=53 ++ DNS_INTROSPECT_PORT=8092 ++ RNDC_KEY=xvysmOR8lnUQRBcunkC6vg== ++ USE_EXTERNAL_TFTP=False ++ ZOOKEEPER_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ ZOOKEEPER_PORT=2181 ++ ZOOKEEPER_PORTS=2888:3888 +++ get_server_list ZOOKEEPER :2181, +++ local server_typ=ZOOKEEPER_NODES +++ local port_with_delim=:2181, +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+=10.0.0.38:2181, +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+=10.0.0.241:2181, +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+=10.0.0.242:2181, +++ '[' -n 10.0.0.38:2181,10.0.0.241:2181,10.0.0.242:2181, ']' +++ echo 10.0.0.38:2181,10.0.0.241:2181,10.0.0.242:2181 ++ ZOOKEEPER_SERVERS=10.0.0.38:2181,10.0.0.241:2181,10.0.0.242:2181 +++ get_server_list ZOOKEEPER ':2181 ' +++ local server_typ=ZOOKEEPER_NODES +++ local 'port_with_delim=:2181 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:2181 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:2181 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:2181 ' +++ '[' -n '10.0.0.38:2181 10.0.0.241:2181 10.0.0.242:2181 ' ']' +++ echo '10.0.0.38:2181 10.0.0.241:2181 10.0.0.242:2181' ++ ZOOKEEPER_SERVERS_SPACE_DELIM='10.0.0.38:2181 10.0.0.241:2181 10.0.0.242:2181' ++ RABBITMQ_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ RABBITMQ_NODE_PORT=5673 +++ get_server_list RABBITMQ :5673, +++ local server_typ=RABBITMQ_NODES +++ local port_with_delim=:5673, +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+=10.0.0.38:5673, +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+=10.0.0.241:5673, +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+=10.0.0.242:5673, +++ '[' -n 10.0.0.38:5673,10.0.0.241:5673,10.0.0.242:5673, ']' +++ echo 10.0.0.38:5673,10.0.0.241:5673,10.0.0.242:5673 ++ RABBITMQ_SERVERS=10.0.0.38:5673,10.0.0.241:5673,10.0.0.242:5673 ++ RABBITMQ_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ RABBITMQ_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ RABBITMQ_SSL_CACERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ RABBITMQ_SSL_FAIL_IF_NO_PEER_CERT=true ++ RABBITMQ_VHOST=/ ++ RABBITMQ_USER=guest ++ RABBITMQ_PASSWORD=guest ++ RABBITMQ_USE_SSL=false ++ RABBITMQ_SSL_VER=tlsv1.2 ++ RABBITMQ_CLIENT_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ RABBITMQ_CLIENT_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ RABBITMQ_CLIENT_SSL_CACERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ RABBITMQ_HEARTBEAT_INTERVAL=60 ++ RABBITMQ_CLUSTER_PARTITION_HANDLING=autoheal ++ RABBITMQ_MIRRORED_QUEUE_MODE=all ++ REDIS_SERVER_PORT=6379 ++ REDIS_SERVER_PASSWORD= +++ get_server_list ANALYTICS ':6379 ' +++ local server_typ=ANALYTICS_NODES +++ local 'port_with_delim=:6379 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:6379 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:6379 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:6379 ' +++ '[' -n '10.0.0.38:6379 10.0.0.241:6379 10.0.0.242:6379 ' ']' +++ echo '10.0.0.38:6379 10.0.0.241:6379 10.0.0.242:6379' ++ REDIS_SERVERS='10.0.0.38:6379 10.0.0.241:6379 10.0.0.242:6379' ++ REDIS_LISTEN_ADDRESS= ++ REDIS_PROTECTED_MODE= ++ REDIS_SSL_ENABLE=false ++ REDIS_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ REDIS_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ REDIS_SSL_CACERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ redis_ssl_config= ++ KAFKA_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ KAFKA_PORT=9092 +++ get_server_list KAFKA ':9092 ' +++ local server_typ=KAFKA_NODES +++ local 'port_with_delim=:9092 ' +++ local server_list= +++ IFS=, +++ read -ra server_list +++ local extended_server_list= +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.38 +++ local server_address=10.0.0.38 +++ extended_server_list+='10.0.0.38:9092 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.241 +++ local server_address=10.0.0.241 +++ extended_server_list+='10.0.0.241:9092 ' +++ for server in '"${server_list[@]}"' ++++ echo 10.0.0.242 +++ local server_address=10.0.0.242 +++ extended_server_list+='10.0.0.242:9092 ' +++ '[' -n '10.0.0.38:9092 10.0.0.241:9092 10.0.0.242:9092 ' ']' +++ echo '10.0.0.38:9092 10.0.0.241:9092 10.0.0.242:9092' ++ KAFKA_SERVERS='10.0.0.38:9092 10.0.0.241:9092 10.0.0.242:9092' ++ KAFKA_SSL_ENABLE=false ++ KAFKA_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ KAFKA_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ KAFKA_SSL_CACERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ KEYSTONE_AUTH_ADMIN_TENANT=admin ++ KEYSTONE_AUTH_ADMIN_USER=admin ++ KEYSTONE_AUTH_ADMIN_PASSWORD=contrail123 ++ KEYSTONE_AUTH_PROJECT_DOMAIN_NAME=Default ++ KEYSTONE_AUTH_USER_DOMAIN_NAME=Default ++ KEYSTONE_AUTH_REGION_NAME=RegionOne ++ KEYSTONE_AUTH_URL_VERSION=/v3 ++ KEYSTONE_AUTH_HOST=10.0.0.38 ++ KEYSTONE_AUTH_PROTO=http ++ KEYSTONE_AUTH_ADMIN_PORT=5000 ++ KEYSTONE_AUTH_PUBLIC_PORT=5000 ++ KEYSTONE_AUTH_URL_TOKENS=/v3/auth/tokens ++ KEYSTONE_AUTH_INSECURE=True ++ KEYSTONE_AUTH_CERTFILE= ++ KEYSTONE_AUTH_KEYFILE= ++ KEYSTONE_AUTH_CA_CERTFILE= ++ KEYSTONE_AUTH_ENDPOINT_TYPE= ++ KEYSTONE_AUTH_SYNC_ON_DEMAND= ++ KEYSTONE_AUTH_INTERFACE=public ++ KUBEMANAGER_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ KUBERNETES_CLUSTER_NAME=k8s ++ KUBERNETES_CNI_META_PLUGIN=multus ++ METADATA_PROXY_SECRET=contrail ++ BARBICAN_TENANT_NAME=service ++ BARBICAN_USER=barbican ++ BARBICAN_PASSWORD=contrail123 ++ AGENT_MODE=kernel ++ EXTERNAL_ROUTERS= ++ SUBCLUSTER= ++ VROUTER_COMPUTE_NODE_ADDRESS= ++ VROUTER_CRYPT_INTERFACE=crypt0 ++ VROUTER_DECRYPT_INTERFACE=decrypt0 ++ VROUTER_DECRYPT_KEY=15 ++ VROUTER_MODULE_OPTIONS= ++ FABRIC_SNAT_HASH_TABLE_SIZE=4096 ++ TSN_EVPN_MODE=False ++ TSN_NODES='[]' ++ PRIORITY_ID= ++ PRIORITY_BANDWIDTH= ++ PRIORITY_SCHEDULING= ++ QOS_QUEUE_ID= ++ QOS_LOGICAL_QUEUES= ++ QOS_DEF_HW_QUEUE=False ++ PRIORITY_TAGGING=True ++ SLO_DESTINATION=collector ++ '[' -n '' ']' ++ SAMPLE_DESTINATION=collector ++ FLOW_EXPORT_RATE=0 ++ WEBUI_NODES=10.0.0.38,10.0.0.241,10.0.0.242 ++ WEBUI_JOB_SERVER_PORT=3000 ++ KUE_UI_PORT=3002 ++ WEBUI_HTTP_LISTEN_PORT=8180 ++ WEBUI_HTTPS_LISTEN_PORT=8143 ++ WEBUI_SSL_KEY_FILE=/etc/contrail/webui_ssl/cs-key.pem ++ WEBUI_SSL_CERT_FILE=/etc/contrail/webui_ssl/cs-cert.pem ++ WEBUI_SSL_CIPHERS=ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-SHA384:ECDHE-RSA-AES256-SHA384:ECDHE-ECDSA-AES128-SHA256:ECDHE-RSA-AES128-SHA256:AES256-SHA ++ WEBUI_STATIC_AUTH_USER=admin ++ WEBUI_STATIC_AUTH_PASSWORD=contrail123 ++ WEBUI_STATIC_AUTH_ROLE=cloudAdmin ++ XMPP_SERVER_PORT=5269 ++ XMPP_SSL_ENABLE=false ++ XMPP_SERVER_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ XMPP_SERVER_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ XMPP_SERVER_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ LINKLOCAL_SERVICE_PORT=80 ++ LINKLOCAL_SERVICE_NAME=metadata ++ LINKLOCAL_SERVICE_IP=169.254.169.254 ++ IPFABRIC_SERVICE_PORT=8775 ++ INTROSPECT_SSL_ENABLE=false ++ INTROSPECT_SSL_INSECURE=True ++ INTROSPECT_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ INTROSPECT_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ INTROSPECT_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ INTROSPECT_LISTEN_ALL=True ++ SANDESH_SSL_ENABLE=false ++ SANDESH_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ SANDESH_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ SANDESH_SERVER_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ SANDESH_SERVER_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ SANDESH_CA_CERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ METADATA_SSL_ENABLE=false ++ METADATA_SSL_CERTFILE= ++ METADATA_SSL_KEYFILE= ++ METADATA_SSL_CA_CERTFILE= ++ METADATA_SSL_CERT_TYPE= ++ CONFIGURE_IPTABLES=false ++ FWAAS_ENABLE=False ++ CONTAINERD_NAMESPACE=k8s.io ++ TOR_AGENT_OVS_KA=10000 ++ TOR_TYPE=ovs ++ TOR_OVS_PROTOCOL=tcp ++ TORAGENT_SSL_CERTFILE=/etc/contrail/ssl/certs/server.pem ++ TORAGENT_SSL_KEYFILE=/etc/contrail/ssl/private/server-privkey.pem ++ TORAGENT_SSL_CACERTFILE=/etc/contrail/ssl/certs/ca-cert.pem ++ [[ /v3 == \/\v\2\.\0 ]] ++ [[ openstack == \o\p\e\n\s\t\a\c\k ]] ++ AUTH_MODE=keystone ++ [[ keystone == \k\e\y\s\t\o\n\e ]] ++ AUTH_PARAMS='--admin_password contrail123' ++ AUTH_PARAMS+=' --admin_tenant_name admin' ++ AUTH_PARAMS+=' --admin_user admin' ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ read -r -d '' sandesh_client_config ++ true ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ xmpp_certs_config= ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ analytics_api_ssl_opts= ++ read -r -d '' rabbitmq_config ++ true ++ read -r -d '' rabbit_config ++ true ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ is_enabled false ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ kafka_ssl_config= ++ [[ -n '' ]] ++ collector_stats_config= ++ [[ -z '' ]] ++ is_enabled False ++ local val=false ++ [[ false == \t\r\u\e ]] ++ [[ false == \y\e\s ]] ++ [[ false == \e\n\a\b\l\e\d ]] ++ export TSN_AGENT_MODE= ++ TSN_AGENT_MODE= ++ [[ -n '' ]] ++ collector_stats_config= ++ [[ -z x ]] ++ RSYSLOGD_XFLOW_LISTEN_PORT=9898 + CONFIG=/etc/cassandra/cassandra.yaml + JVM_OPTIONS_CONFIG=/etc/cassandra/jvm.options + cp /etc/cassandra/cassandra.origin /etc/cassandra/cassandra.yaml + cp /etc/cassandra/jvm.options.origin /etc/cassandra/jvm.options + for i in '{1..10}' ++ cut -d ' ' -f 1 ++ find_my_ip_and_order_for_node_list 10.0.0.38,10.0.0.241,10.0.0.242 ++ local servers=10.0.0.38,10.0.0.241,10.0.0.242 ++ local server_list= ++ IFS=, ++ read -ra server_list +++ get_local_ips +++ cat /proc/net/fib_trie +++ tr '\n' , +++ grep -vi host +++ awk '/32 host/ { print f } {f=$2}' +++ sort +++ uniq ++ local local_ips=,10.0.0.242,10.20.0.252,127.0.0.1,172.17.0.1,, ++ local ord=1 ++ for server in '"${server_list[@]}"' ++ local ret=0 +++ python3 -c 'import socket; print(socket.gethostbyname('\''10.0.0.38'\''))' ++ local server_ip=10.0.0.38 ++ [[ 0 == 0 ]] ++ [[ -n 10.0.0.38 ]] ++ [[ ,10.0.0.242,10.20.0.252,127.0.0.1,172.17.0.1,, =~ ,10\.0\.0\.38, ]] ++ (( ord+=1 )) ++ for server in '"${server_list[@]}"' ++ local ret=0 +++ python3 -c 'import socket; print(socket.gethostbyname('\''10.0.0.241'\''))' ++ local server_ip=10.0.0.241 ++ [[ 0 == 0 ]] ++ [[ -n 10.0.0.241 ]] ++ [[ ,10.0.0.242,10.20.0.252,127.0.0.1,172.17.0.1,, =~ ,10\.0\.0\.241, ]] ++ (( ord+=1 )) ++ for server in '"${server_list[@]}"' ++ local ret=0 +++ python3 -c 'import socket; print(socket.gethostbyname('\''10.0.0.242'\''))' ++ local server_ip=10.0.0.242 ++ [[ 0 == 0 ]] ++ [[ -n 10.0.0.242 ]] ++ [[ ,10.0.0.242,10.20.0.252,127.0.0.1,172.17.0.1,, =~ ,10\.0\.0\.242, ]] ++ echo 10.0.0.242 3 ++ return + my_ip=10.0.0.242 + '[' -n 10.0.0.242 ']' + break + '[' -z 10.0.0.242 ']' ++ echo 10.0.0.38,10.0.0.241,10.0.0.242 ++ tr , ' ' ++ wc -w + export CASSANDRA_COUNT=3 + CASSANDRA_COUNT=3 ++ echo 10.0.0.38,10.0.0.241,10.0.0.242 ++ sed 's/,/", "/g' + export 'CASSANDRA_CONNECT_POINTS=10.0.0.38", "10.0.0.241", "10.0.0.242' + CASSANDRA_CONNECT_POINTS='10.0.0.38", "10.0.0.241", "10.0.0.242' ++ echo 10.0.0.38,10.0.0.241,10.0.0.242 ++ cut -d , -f 1,2 + export CASSANDRA_SEEDS=10.0.0.38,10.0.0.241 + CASSANDRA_SEEDS=10.0.0.38,10.0.0.241 + export CASSANDRA_LISTEN_ADDRESS=10.0.0.242 + CASSANDRA_LISTEN_ADDRESS=10.0.0.242 + export CASSANDRA_RPC_ADDRESS=10.0.0.242 + CASSANDRA_RPC_ADDRESS=10.0.0.242 + echo 'INFO: JVM_EXTRA_OPTS=-Xms1g -Xmx2g' INFO: JVM_EXTRA_OPTS=-Xms1g -Xmx2g + for yaml in Xmx Xms ++ echo -Xms1g -Xmx2g ++ sed -n 's/.*\(-Xmx[0-9]*[mMgG]\).*/\1/p' + opt=-Xmx2g + [[ -n -Xmx2g ]] ++ echo -Xms1g -Xmx2g ++ sed 's/-Xmx[0-9]*[mMgG]//g' + JVM_EXTRA_OPTS='-Xms1g ' + sed -i 's/^[#]*-Xmx.*/-Xmx2g/g' /etc/cassandra/jvm.options + for yaml in Xmx Xms ++ sed -n 's/.*\(-Xms[0-9]*[mMgG]\).*/\1/p' ++ echo -Xms1g + opt=-Xms1g + [[ -n -Xms1g ]] ++ echo -Xms1g ++ sed 's/-Xms[0-9]*[mMgG]//g' + JVM_EXTRA_OPTS= + sed -i 's/^[#]*-Xms.*/-Xms1g/g' /etc/cassandra/jvm.options + export 'JVM_EXTRA_OPTS= -Dcassandra.rpc_port=9161 -Dcassandra.native_transport_port=9041 -Dcassandra.ssl_storage_port=7013 -Dcassandra.storage_port=7012 -Dcassandra.jmx.local.port=7201' + JVM_EXTRA_OPTS=' -Dcassandra.rpc_port=9161 -Dcassandra.native_transport_port=9041 -Dcassandra.ssl_storage_port=7013 -Dcassandra.storage_port=7012 -Dcassandra.jmx.local.port=7201' + is_enabled true + local val=true + [[ true == \t\r\u\e ]] + export LOCAL_JMX=no + LOCAL_JMX=no + export 'JVM_EXTRA_OPTS= -Dcassandra.rpc_port=9161 -Dcassandra.native_transport_port=9041 -Dcassandra.ssl_storage_port=7013 -Dcassandra.storage_port=7012 -Dcassandra.jmx.local.port=7201 -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access -Dcassandra.jmx.remote.port=7201 -Dcom.sun.management.jmxremote.rmi.port=7201' + JVM_EXTRA_OPTS=' -Dcassandra.rpc_port=9161 -Dcassandra.native_transport_port=9041 -Dcassandra.ssl_storage_port=7013 -Dcassandra.storage_port=7012 -Dcassandra.jmx.local.port=7201 -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access -Dcassandra.jmx.remote.port=7201 -Dcom.sun.management.jmxremote.rmi.port=7201' + is_enabled false + local val=false + [[ false == \t\r\u\e ]] + [[ false == \y\e\s ]] + [[ false == \e\n\a\b\l\e\d ]] + cat + change_variable memtable_flush_writers 4 + local VARIABLE_NAME=memtable_flush_writers + local VARIABLE_VALUE=4 + sed -i 's/.*\(memtable_flush_writers\):.*\([0-9a-z]\)/\1: 4/g' /etc/cassandra/cassandra.yaml + change_variable concurrent_compactors 4 + local VARIABLE_NAME=concurrent_compactors + local VARIABLE_VALUE=4 + sed -i 's/.*\(concurrent_compactors\):.*\([0-9a-z]\)/\1: 4/g' /etc/cassandra/cassandra.yaml + change_variable compaction_throughput_mb_per_sec 256 + local VARIABLE_NAME=compaction_throughput_mb_per_sec + local VARIABLE_VALUE=256 + sed -i 's/.*\(compaction_throughput_mb_per_sec\):.*\([0-9a-z]\)/\1: 256/g' /etc/cassandra/cassandra.yaml + change_variable concurrent_reads 64 + local VARIABLE_NAME=concurrent_reads + local VARIABLE_VALUE=64 + sed -i 's/.*\(concurrent_reads\):.*\([0-9a-z]\)/\1: 64/g' /etc/cassandra/cassandra.yaml + change_variable concurrent_writes 64 + local VARIABLE_NAME=concurrent_writes + local VARIABLE_VALUE=64 + sed -i 's/.*\(concurrent_writes\):.*\([0-9a-z]\)/\1: 64/g' /etc/cassandra/cassandra.yaml + change_variable memtable_allocation_type offheap_objects + local VARIABLE_NAME=memtable_allocation_type + local VARIABLE_VALUE=offheap_objects + sed -i 's/.*\(memtable_allocation_type\):.*\([0-9a-z]\)/\1: offheap_objects/g' /etc/cassandra/cassandra.yaml + log_levels_map=([SYS_DEBUG]='DEBUG' [SYS_INFO]='INFO' [SYS_NOTICE]='INFO' [SYS_ERROR]="ERROR") + declare -A log_levels_map + log_level=DEBUG + '[' -n DEBUG ']' + sed -i 's/\(; cluster_name=contrail_database; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=stop; commitlog_compression=null; commitlog_directory=/var/lib/cassandra/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=256; concurrent_compactors=4; concurrent_counter_writes=32; concurrent_materialized_view_writes=32; concurrent_reads=64; concurrent_replicates=null; concurrent_writes=64; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=2000; cross_node_timeout=false; data_file_directories=[Ljava.lang.String;@6b19b79; disk_access_mode=auto; disk_failure_policy=stop; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=true; dynamic_snitch_badness_threshold=0.1; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_materialized_views=true; enable_scripted_user_defined_functions=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; encryption_options=null; endpoint_snitch=SimpleSnitch; file_cache_round_up=null; file_cache_size_in_mb=null; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=null; hints_flush_period_in_ms=10000; incremental_backups=false; index_interval=null; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_authenticator=null; internode_compression=dc; internode_recv_buff_size_in_bytes=0; internode_send_buff_size_in_bytes=0; key_cache_keys_to_save=2147483647; key_cache_save_period=14400; key_cache_size_in_mb=null; listen_address=10.0.0.242; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=offheap_objects; memtable_cleanup_threshold=null; memtable_flush_writers=4; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; num_tokens=256; otc_backlog_expiration_interval_ms=200; otc_coalescing_enough_coalesced_messages=8; otc_coalescing_strategy=DISABLED; otc_coalescing_window_us=200; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=2000; phi_convict_threshold=8.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; read_request_timeout_in_ms=5000; request_scheduler=org.apache.cassandra.scheduler.NoScheduler; request_scheduler_id=null; request_scheduler_options=null; request_timeout_in_ms=10000; role_manager=CassandraRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=2000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=10.0.0.242; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; rpc_listen_backlog=50; rpc_max_threads=2147483647; rpc_min_threads=16; rpc_port=9160; rpc_recv_buff_size_in_bytes=null; rpc_send_buff_size_in_bytes=null; rpc_server_type=sync; saved_caches_directory=/var/lib/cassandra/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=10.0.0.38,10.0.0.241}; server_encryption_options=; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; start_rpc=true; storage_port=7000; stream_throughput_outbound_megabits_per_sec=200; streaming_keep_alive_period_in_secs=300; streaming_socket_timeout_in_ms=86400000; thrift_framed_transport_size_in_mb=15; thrift_max_message_length_in_mb=16; thrift_prepared_statements_cache_size_mb=null; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions@2a32de6c; trickle_fsync=false; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; windows_timer_interval=1; write_request_timeout_in_ms=2000] INFO [main] 2025-10-16 05:22:34,086 DatabaseDescriptor.java:367 - DiskAccessMode 'auto' determined to be mmap, indexAccessMode is mmap INFO [main] 2025-10-16 05:22:34,087 DatabaseDescriptor.java:425 - Global memtable on-heap threshold is enabled at 502MB INFO [main] 2025-10-16 05:22:34,087 DatabaseDescriptor.java:429 - Global memtable off-heap threshold is enabled at 502MB INFO [main] 2025-10-16 05:22:34,113 RateBasedBackPressure.java:123 - Initialized back-pressure with high ratio: 0.9, factor: 5, flow: FAST, window size: 2000. INFO [main] 2025-10-16 05:22:34,114 DatabaseDescriptor.java:729 - Back-pressure is disabled with strategy org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}. INFO [main] 2025-10-16 05:22:34,267 JMXServerUtils.java:246 - Configured JMX server at: service:jmx:rmi://0.0.0.0/jndi/rmi://0.0.0.0:7201/jmxrmi INFO [main] 2025-10-16 05:22:34,270 CassandraDaemon.java:473 - Hostname: cn-jenkins-deploy-platform-ansible-os-4340-3. INFO [main] 2025-10-16 05:22:34,271 CassandraDaemon.java:480 - JVM vendor/version: OpenJDK 64-Bit Server VM/1.8.0_322 INFO [main] 2025-10-16 05:22:34,273 CassandraDaemon.java:481 - Heap size: 984.000MiB/1.961GiB INFO [main] 2025-10-16 05:22:34,274 CassandraDaemon.java:486 - Code Cache Non-heap memory: init = 2555904(2496K) used = 4389568(4286K) committed = 4456448(4352K) max = 251658240(245760K) INFO [main] 2025-10-16 05:22:34,274 CassandraDaemon.java:486 - Metaspace Non-heap memory: init = 0(0K) used = 18806920(18366K) committed = 19398656(18944K) max = -1(-1K) INFO [main] 2025-10-16 05:22:34,274 CassandraDaemon.java:486 - Compressed Class Space Non-heap memory: init = 0(0K) used = 2197624(2146K) committed = 2359296(2304K) max = 1073741824(1048576K) INFO [main] 2025-10-16 05:22:34,275 CassandraDaemon.java:486 - Par Eden Space Heap memory: init = 335544320(327680K) used = 87277544(85231K) committed = 335544320(327680K) max = 335544320(327680K) INFO [main] 2025-10-16 05:22:34,275 CassandraDaemon.java:486 - Par Survivor Space Heap memory: init = 41943040(40960K) used = 0(0K) committed = 41943040(40960K) max = 41943040(40960K) INFO [main] 2025-10-16 05:22:34,275 CassandraDaemon.java:486 - CMS Old Gen Heap memory: init = 654311424(638976K) used = 0(0K) committed = 654311424(638976K) max = 1728053248(1687552K) INFO [main] 2025-10-16 05:22:34,276 CassandraDaemon.java:488 - Classpath: /opt/cassandra/conf:/opt/cassandra/build/classes/main:/opt/cassandra/build/classes/thrift:/opt/cassandra/lib/airline-0.6.jar:/opt/cassandra/lib/antlr-runtime-3.5.2.jar:/opt/cassandra/lib/apache-cassandra-3.11.3.jar:/opt/cassandra/lib/apache-cassandra-thrift-3.11.3.jar:/opt/cassandra/lib/asm-5.0.4.jar:/opt/cassandra/lib/caffeine-2.2.6.jar:/opt/cassandra/lib/cassandra-driver-core-3.0.1-shaded.jar:/opt/cassandra/lib/commons-cli-1.1.jar:/opt/cassandra/lib/commons-codec-1.9.jar:/opt/cassandra/lib/commons-lang3-3.1.jar:/opt/cassandra/lib/commons-math3-3.2.jar:/opt/cassandra/lib/compress-lzf-0.8.4.jar:/opt/cassandra/lib/concurrentlinkedhashmap-lru-1.4.jar:/opt/cassandra/lib/concurrent-trees-2.4.0.jar:/opt/cassandra/lib/disruptor-3.0.1.jar:/opt/cassandra/lib/ecj-4.4.2.jar:/opt/cassandra/lib/guava-18.0.jar:/opt/cassandra/lib/HdrHistogram-2.1.9.jar:/opt/cassandra/lib/high-scale-lib-1.0.6.jar:/opt/cassandra/lib/hppc-0.5.4.jar:/opt/cassandra/lib/jackson-core-asl-1.9.13.jar:/opt/cassandra/lib/jackson-mapper-asl-1.9.13.jar:/opt/cassandra/lib/jamm-0.3.0.jar:/opt/cassandra/lib/javax.inject.jar:/opt/cassandra/lib/jbcrypt-0.3m.jar:/opt/cassandra/lib/jcl-over-slf4j-1.7.7.jar:/opt/cassandra/lib/jctools-core-1.2.1.jar:/opt/cassandra/lib/jflex-1.6.0.jar:/opt/cassandra/lib/jna-4.2.2.jar:/opt/cassandra/lib/joda-time-2.4.jar:/opt/cassandra/lib/json-simple-1.1.jar:/opt/cassandra/lib/jstackjunit-0.0.1.jar:/opt/cassandra/lib/libthrift-0.13.0.jar:/opt/cassandra/lib/log4j-over-slf4j-1.7.7.jar:/opt/cassandra/lib/logback-classic-1.2.9.jar:/opt/cassandra/lib/logback-core-1.2.9.jar:/opt/cassandra/lib/lz4-1.3.0.jar:/opt/cassandra/lib/metrics-core-3.1.5.jar:/opt/cassandra/lib/metrics-jvm-3.1.5.jar:/opt/cassandra/lib/metrics-logback-3.1.5.jar:/opt/cassandra/lib/netty-all-4.1.39.Final.jar:/opt/cassandra/lib/ohc-core-0.4.4.jar:/opt/cassandra/lib/ohc-core-j8-0.4.4.jar:/opt/cassandra/lib/reporter-config3-3.0.3.jar:/opt/cassandra/lib/reporter-config-base-3.0.3.jar:/opt/cassandra/lib/sigar-1.6.4.jar:/opt/cassandra/lib/slf4j-api-1.7.7.jar:/opt/cassandra/lib/snakeyaml-1.11.jar:/opt/cassandra/lib/snappy-java-1.1.1.7.jar:/opt/cassandra/lib/snowball-stemmer-1.3.0.581.1.jar:/opt/cassandra/lib/ST4-4.0.8.jar:/opt/cassandra/lib/stream-2.5.2.jar:/opt/cassandra/lib/thrift-server-0.3.7.jar:/opt/cassandra/lib/jsr223/*/*.jar:/opt/cassandra/lib/jamm-0.3.0.jar INFO [main] 2025-10-16 05:22:34,276 CassandraDaemon.java:490 - JVM Arguments: [-Xloggc:/opt/cassandra/logs/gc.log, -ea, -XX:+UseThreadPriorities, -XX:ThreadPriorityPolicy=42, -XX:+HeapDumpOnOutOfMemoryError, -Xss256k, -XX:StringTableSize=1000003, -XX:+AlwaysPreTouch, -XX:-UseBiasedLocking, -XX:+UseTLAB, -XX:+ResizeTLAB, -XX:+UseNUMA, -XX:+PerfDisableSharedMem, -Djava.net.preferIPv4Stack=true, -Xms1g, -Xmx2g, -XX:+UseParNewGC, -XX:+UseConcMarkSweepGC, -XX:+CMSParallelRemarkEnabled, -XX:SurvivorRatio=8, -XX:MaxTenuringThreshold=1, -XX:CMSInitiatingOccupancyFraction=75, -XX:+UseCMSInitiatingOccupancyOnly, -XX:CMSWaitDuration=10000, -XX:+CMSParallelInitialMarkEnabled, -XX:+CMSEdenChunksRecordAlways, -XX:+CMSClassUnloadingEnabled, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -XX:+PrintHeapAtGC, -XX:+PrintTenuringDistribution, -XX:+PrintGCApplicationStoppedTime, -XX:+PrintPromotionFailure, -XX:+UseGCLogFileRotation, -XX:NumberOfGCLogFiles=10, -XX:GCLogFileSize=10M, -Xmn400M, -XX:+UseCondCardMark, -XX:CompileCommandFile=/opt/cassandra/conf/hotspot_compiler, -javaagent:/opt/cassandra/lib/jamm-0.3.0.jar, -Dcassandra.jmx.remote.port=7199, -Dcom.sun.management.jmxremote.rmi.port=7199, -Dcom.sun.management.jmxremote.authenticate=true, -Dcom.sun.management.jmxremote.password.file=/etc/cassandra/jmxremote.password, -Djava.library.path=/opt/cassandra/lib/sigar-bin, -Dcassandra.rpc_port=9161, -Dcassandra.native_transport_port=9041, -Dcassandra.ssl_storage_port=7013, -Dcassandra.storage_port=7012, -Dcassandra.jmx.local.port=7201, -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access, -Dcassandra.jmx.remote.port=7201, -Dcom.sun.management.jmxremote.rmi.port=7201, -Dcassandra.libjemalloc=/usr/lib64/libjemalloc.so.1, -XX:OnOutOfMemoryError=kill -9 %p, -Dlogback.configurationFile=logback.xml, -Dcassandra.logdir=/opt/cassandra/logs, -Dcassandra.storagedir=/opt/cassandra/data, -Dcassandra-foreground=yes] WARN [main] 2025-10-16 05:22:34,318 NativeLibrary.java:187 - Unable to lock JVM memory (ENOMEM). This can result in part of the JVM being swapped out, especially with mmapped I/O enabled. Increase RLIMIT_MEMLOCK or run Cassandra as root. INFO [main] 2025-10-16 05:22:34,318 StartupChecks.java:140 - jemalloc seems to be preloaded from /usr/lib64/libjemalloc.so.1 INFO [main] 2025-10-16 05:22:34,318 StartupChecks.java:176 - JMX is enabled to receive remote connections on port: 7201 INFO [main] 2025-10-16 05:22:34,319 SigarLibrary.java:44 - Initializing SIGAR library INFO [main] 2025-10-16 05:22:34,328 SigarLibrary.java:180 - Checked OS settings and found them configured for optimal performance. WARN [main] 2025-10-16 05:22:34,337 StartupChecks.java:311 - Maximum number of memory map areas per process (vm.max_map_count) 128960 is too low, recommended value: 1048575, you can change it with sysctl. WARN [main] 2025-10-16 05:22:34,350 StartupChecks.java:332 - Directory /var/lib/cassandra/commitlog doesn't exist WARN [main] 2025-10-16 05:22:34,351 StartupChecks.java:332 - Directory /var/lib/cassandra/saved_caches doesn't exist WARN [main] 2025-10-16 05:22:34,352 StartupChecks.java:332 - Directory /opt/cassandra/data/hints doesn't exist INFO [main] 2025-10-16 05:22:34,410 QueryProcessor.java:116 - Initialized prepared statement caches with 10 MB (native) and 10 MB (Thrift) INFO [main] 2025-10-16 05:22:34,987 ColumnFamilyStore.java:411 - Initializing system.IndexInfo INFO [main] 2025-10-16 05:22:36,129 ColumnFamilyStore.java:411 - Initializing system.batches INFO [main] 2025-10-16 05:22:36,162 ColumnFamilyStore.java:411 - Initializing system.paxos INFO [main] 2025-10-16 05:22:36,187 ColumnFamilyStore.java:411 - Initializing system.local INFO [main] 2025-10-16 05:22:36,193 ColumnFamilyStore.java:411 - Initializing system.peers INFO [main] 2025-10-16 05:22:36,202 ColumnFamilyStore.java:411 - Initializing system.peer_events INFO [main] 2025-10-16 05:22:36,207 ColumnFamilyStore.java:411 - Initializing system.range_xfers INFO [main] 2025-10-16 05:22:36,211 ColumnFamilyStore.java:411 - Initializing system.compaction_history INFO [main] 2025-10-16 05:22:36,220 ColumnFamilyStore.java:411 - Initializing system.sstable_activity INFO [main] 2025-10-16 05:22:36,224 ColumnFamilyStore.java:411 - Initializing system.size_estimates INFO [main] 2025-10-16 05:22:36,227 ColumnFamilyStore.java:411 - Initializing system.available_ranges INFO [main] 2025-10-16 05:22:36,231 ColumnFamilyStore.java:411 - Initializing system.transferred_ranges INFO [main] 2025-10-16 05:22:36,234 ColumnFamilyStore.java:411 - Initializing system.views_builds_in_progress INFO [main] 2025-10-16 05:22:36,237 ColumnFamilyStore.java:411 - Initializing system.built_views INFO [main] 2025-10-16 05:22:36,240 ColumnFamilyStore.java:411 - Initializing system.hints INFO [main] 2025-10-16 05:22:36,243 ColumnFamilyStore.java:411 - Initializing system.batchlog INFO [main] 2025-10-16 05:22:36,252 ColumnFamilyStore.java:411 - Initializing system.prepared_statements INFO [main] 2025-10-16 05:22:36,255 ColumnFamilyStore.java:411 - Initializing system.schema_keyspaces INFO [main] 2025-10-16 05:22:36,258 ColumnFamilyStore.java:411 - Initializing system.schema_columnfamilies INFO [main] 2025-10-16 05:22:36,261 ColumnFamilyStore.java:411 - Initializing system.schema_columns INFO [main] 2025-10-16 05:22:36,264 ColumnFamilyStore.java:411 - Initializing system.schema_triggers INFO [main] 2025-10-16 05:22:36,267 ColumnFamilyStore.java:411 - Initializing system.schema_usertypes INFO [main] 2025-10-16 05:22:36,289 ColumnFamilyStore.java:411 - Initializing system.schema_functions INFO [main] 2025-10-16 05:22:36,305 ColumnFamilyStore.java:411 - Initializing system.schema_aggregates INFO [main] 2025-10-16 05:22:36,307 ViewManager.java:137 - Not submitting build tasks for views in keyspace system as storage service is not initialized INFO [main] 2025-10-16 05:22:36,457 ApproximateTime.java:44 - Scheduling approximate time-check task with a precision of 10 milliseconds INFO [main] 2025-10-16 05:22:36,499 ColumnFamilyStore.java:411 - Initializing system_schema.keyspaces INFO [main] 2025-10-16 05:22:36,503 ColumnFamilyStore.java:411 - Initializing system_schema.tables INFO [main] 2025-10-16 05:22:36,508 ColumnFamilyStore.java:411 - Initializing system_schema.columns INFO [main] 2025-10-16 05:22:36,511 ColumnFamilyStore.java:411 - Initializing system_schema.triggers INFO [main] 2025-10-16 05:22:36,514 ColumnFamilyStore.java:411 - Initializing system_schema.dropped_columns INFO [main] 2025-10-16 05:22:36,517 ColumnFamilyStore.java:411 - Initializing system_schema.views INFO [main] 2025-10-16 05:22:36,520 ColumnFamilyStore.java:411 - Initializing system_schema.types INFO [main] 2025-10-16 05:22:36,522 ColumnFamilyStore.java:411 - Initializing system_schema.functions INFO [main] 2025-10-16 05:22:36,538 ColumnFamilyStore.java:411 - Initializing system_schema.aggregates INFO [main] 2025-10-16 05:22:36,569 ColumnFamilyStore.java:411 - Initializing system_schema.indexes INFO [main] 2025-10-16 05:22:36,571 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_schema as storage service is not initialized INFO [MemtableFlushWriter:1] 2025-10-16 05:22:37,834 CacheService.java:112 - Initializing key cache with capacity of 49 MBs. INFO [MemtableFlushWriter:1] 2025-10-16 05:22:37,841 CacheService.java:134 - Initializing row cache with capacity of 0 MBs INFO [MemtableFlushWriter:1] 2025-10-16 05:22:37,859 CacheService.java:163 - Initializing counter cache with capacity of 24 MBs INFO [MemtableFlushWriter:1] 2025-10-16 05:22:37,860 CacheService.java:174 - Scheduling counter cache save to every 7200 seconds (going to save all keys). INFO [CompactionExecutor:4] 2025-10-16 05:22:38,266 BufferPool.java:230 - Global buffer pool is enabled, when pool is exhausted (max is 502.000MiB) it will allocate on heap INFO [main] 2025-10-16 05:22:38,405 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-10-16 05:22:38,478 StorageService.java:607 - Token metadata: INFO [pool-4-thread-1] 2025-10-16 05:22:38,513 AutoSavingCache.java:174 - Completed loading (1 ms; 4 keys) KeyCache cache INFO [main] 2025-10-16 05:22:38,582 CommitLog.java:152 - No commitlog files found; skipping replay INFO [main] 2025-10-16 05:22:38,592 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-10-16 05:22:38,676 StorageService.java:607 - Token metadata: INFO [main] 2025-10-16 05:22:38,813 QueryProcessor.java:163 - Preloaded 0 prepared statements INFO [main] 2025-10-16 05:22:38,814 StorageService.java:618 - Cassandra version: 3.11.3 INFO [main] 2025-10-16 05:22:38,814 StorageService.java:619 - Thrift API version: 20.1.0 INFO [main] 2025-10-16 05:22:38,815 StorageService.java:620 - CQL supported versions: 3.4.4 (default: 3.4.4) INFO [main] 2025-10-16 05:22:38,815 StorageService.java:622 - Native protocol supported versions: 3/v3, 4/v4, 5/v5-beta (default: 4/v4) INFO [main] 2025-10-16 05:22:38,867 IndexSummaryManager.java:85 - Initializing index summary manager with a memory pool size of 49 MB and a resize interval of 60 minutes INFO [main] 2025-10-16 05:22:38,877 MessagingService.java:761 - Starting Messaging Service on /10.0.0.242:7012 (ens3) WARN [main] 2025-10-16 05:22:38,885 SystemKeyspace.java:1087 - No host ID found, created 098e411d-a37d-4046-8678-fa568ca407df (Note: This should happen exactly once per node). INFO [main] 2025-10-16 05:22:38,907 OutboundTcpConnection.java:108 - OutboundTcpConnection using coalescing strategy DISABLED INFO [HANDSHAKE-/10.0.0.241] 2025-10-16 05:22:39,955 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.241 INFO [HANDSHAKE-/10.0.0.38] 2025-10-16 05:22:41,570 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 INFO [main] 2025-10-16 05:22:41,955 StorageService.java:704 - Loading persisted ring state INFO [main] 2025-10-16 05:22:41,956 StorageService.java:822 - Starting up server gossip INFO [main] 2025-10-16 05:22:42,073 StorageService.java:1446 - JOINING: waiting for ring information INFO [HANDSHAKE-/10.0.0.38] 2025-10-16 05:22:43,006 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 INFO [MigrationStage:1] 2025-10-16 05:22:43,906 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_distributed as storage service is not initialized INFO [MigrationStage:1] 2025-10-16 05:22:43,909 ColumnFamilyStore.java:411 - Initializing system_distributed.parent_repair_history INFO [MigrationStage:1] 2025-10-16 05:22:43,913 ColumnFamilyStore.java:411 - Initializing system_distributed.repair_history INFO [MigrationStage:1] 2025-10-16 05:22:43,916 ColumnFamilyStore.java:411 - Initializing system_distributed.view_build_status INFO [HANDSHAKE-/10.0.0.241] 2025-10-16 05:22:43,924 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.241 INFO [main] 2025-10-16 05:22:44,075 StorageService.java:1446 - JOINING: schema complete, ready to bootstrap INFO [main] 2025-10-16 05:22:44,075 StorageService.java:1446 - JOINING: waiting for pending range calculation INFO [main] 2025-10-16 05:22:44,076 StorageService.java:1446 - JOINING: calculation complete, ready to bootstrap INFO [main] 2025-10-16 05:22:44,077 StorageService.java:1446 - JOINING: getting bootstrap token INFO [main] 2025-10-16 05:22:44,078 BootStrapper.java:228 - Generated random tokens. tokens are [3865232099155310382, 5952851990590385243, 3240563202895670566, -1148421334932770250, -8225122346858646967, -2981274944881219572, -1122485615342174429, 4606393749740050827, 8476857092308191734, 3030588890644931510, -7219270434980484716, 5070364096661557868, 6466035134996645299, 7926987489270709243, -2859954181064986317, -4919629009555940254, -7640289567122276749, 8809828380506970146, 501522118087435457, 4384624772220997031, -1234469585154011053, -5601688050122438173, 1569218411336921228, 5237822838403033083, -4366652450095138937, 6480545411255989452, 233724486912859643, -4764034114964366773, 3695664036453928894, 8586368560238158908, 2892009789533283227, 7739613978207955714, -5526012803853879894, 4673046324942525286, 4180112473064902229, -184907261125272379, -675858733421903165, 4586218657462390316, 2769766058784149877, -1330952702924886258, -1715951035943847558, -546841866581489640, 1264796929087127800, -9076589835447936054, 4466311374397751910, -7678525806989887006, -2221599587513302893, 1673436004818470392, -3821730512769182254, 1525780980035280114, 5023295557663276336, 2836865183015497635, -3741224166047815792, -4234629247318450072, 779103728941929168, 1452742733987679947, 5847572623598330087, -1754398647267424933, 1023710574378442169, 3673474662376195433, 5543463606295835332, -392029869347299282, 4234162621617363577, -4813791488509643007, -3531636344889451511, 4936499777192508733, -7390836222643716542, 604780858942883370, -8385375980107047306, -6469095515862835779, 1957968757243793247, 3566269909373592856, -775307473100612482, 8585244511082622246, 7017783984413538767, -5120939234933750277, -597444285261257392, 6361256506469105905, 5235013197166584679, 2335198889157645778, 615842487493102065, -4629516826426624893, 3336554095973842127, -3293513670911660473, 2569633045331780282, -5999191512721443910, 77728647517695064, 8661073688073686817, -1873273496814937553, -5915571320365946764, -5815020260733370025, -7050290616687781989, 1221407080455191471, 8755057967890978299, -3611605159394968542, -2637700396918588570, 9033631800095651322, 7991106925625511207, 3960180770604945690, 7484953508503956574, 3157363434991878462, 3814921644584286801, 6910117400858066031, -5802498387193985971, -3824407340925453230, 5964198075320259850, 3246009993620776454, -514528772324163016, 6114558767938749951, 2882944036341699070, 2834965570769247106, -8345448320038572402, -2580223028720700975, -7905385344904519900, 7019506245055488793, 4181380819762484673, 216518570387412129, -295797731683704018, 7621886335007657630, -3653167035823285031, -4054503488734300494, -4536787329871394421, 1136683895312721610, 5026660684770498137, -7636523598898472952, -5703688183224259923, -5893416644393625715, 4702056683091557238, 712252187895585143, -7044172591420805961, 7971746203162708058, 6577035847637701233, -6963335994682042913, -2736979692793242142, -8918432740622949269, 1094968173268150266, 811453690600884590, -5726635861736788057, 2827774050137744428, 3275495825171781855, -4688858860399159876, -6676712882201264545, 7183478296953351798, -5747802695508707345, -7575920852765325395, -6398848602890001912, 995423515595922102, -402284296954540900, 7919135024846827983, 5773863082162895615, -4053252513129157340, 3991234581250826323, -1370510349670207173, -7003423919442890025, -2763372511037094326, 2546419350843339451, 1277973462797918947, 6815038078903172967, -8636126789088298431, 6322120371412670760, -655050326407568157, -3756604564210549745, -6293227713052197456, -190413793917570430, -2081873416720215450, 3528462649820529262, -684089748246519937, 7815233160955215235, -3236268020331366655, -2931450487814623647, 553397566181718291, 8683397936265121232, -873162731219207902, -3205885714676090598, 9030772085880197433, 855468030802882975, 7972973835592095602, 6455192313484128761, 6244809833295487994, -1566990106862673860, 6134443290042001001, -9209456500529342216, 5266390284938709008, -4803469450215782522, -7568208296246460626, 7327945051951642828, 6646375540639311784, 3103891783707885182, -6006848700026616754, 3746987514502335585, 3715361375216927128, 678099646916452557, 4095692096309391384, -1189642194835509329, 570885851432215747, 5494730972676586810, 4434161795058572617, -239058762841958178, -6091473887186623076, 3515279494042486721, 1431325337424230325, 154711859583290675, -6629033740889623217, 7892624676486468408, -6494573418728971940, 8959286372365191533, 2205869249560139165, 6015529335334559334, -8656051150893125007, -2976423085650191513, 6690376229882708260, 6434825285304077605, 5530481021300547055, 301852749886943130, -7305657642340971135, 3220767225740491228, 2968386505189148816, -4690695199001246429, -5249239732791944507, 7993199621543057712, -5747671335129439950, 3853143579874932842, 3964956702950941844, -8996780021335254741, 2217077795969506215, -6082264198933757746, 3752459118850308943, 2144710760865015885, 3664305369054715477, -3779375561136824309, -6964447335695006600, 6478721075999572362, 8227225588450768242, 1302080116720502536, 5029978552281406544, 4282086647115638124, -3745451879711745807, 604911619251529133, 6976327560632823693, 5110901750269140395, 6664447748659515253, -3151652423972229931, 7756793064548286700, 6832532424863132016, 4679568919244435889, 8423411312274512854, -45980178796501036, 4330917489463179661, 7624533946582132459, -7181361911088327991, -2224491597533857954, 3060562203777787315, 5340803614128372302, 3066023252551514619, -3080319591722426961, 5281792349889613330] INFO [main] 2025-10-16 05:22:44,139 StorageService.java:1446 - JOINING: sleeping 30000 ms for pending range setup INFO [GossipStage:1] 2025-10-16 05:22:44,737 Gossiper.java:1055 - Node /10.0.0.241 is now part of the cluster INFO [RequestResponseStage-6] 2025-10-16 05:22:44,815 Gossiper.java:1019 - InetAddress /10.0.0.241 is now UP INFO [GossipStage:1] 2025-10-16 05:22:44,839 TokenMetadata.java:479 - Updating topology for /10.0.0.241 INFO [GossipStage:1] 2025-10-16 05:22:44,839 TokenMetadata.java:479 - Updating topology for /10.0.0.241 INFO [GossipStage:1] 2025-10-16 05:22:44,840 Gossiper.java:1055 - Node /10.0.0.38 is now part of the cluster INFO [RequestResponseStage-1] 2025-10-16 05:22:44,844 Gossiper.java:1019 - InetAddress /10.0.0.38 is now UP WARN [GossipTasks:1] 2025-10-16 05:22:44,999 FailureDetector.java:288 - Not marking nodes down due to local pause of 6503581728 > 5000000000 INFO [MigrationStage:1] 2025-10-16 05:22:45,071 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_auth as storage service is not initialized INFO [MigrationStage:1] 2025-10-16 05:22:45,074 ColumnFamilyStore.java:411 - Initializing system_auth.resource_role_permissons_index INFO [MigrationStage:1] 2025-10-16 05:22:45,078 ColumnFamilyStore.java:411 - Initializing system_auth.role_members INFO [MigrationStage:1] 2025-10-16 05:22:45,082 ColumnFamilyStore.java:411 - Initializing system_auth.role_permissions INFO [MigrationStage:1] 2025-10-16 05:22:45,084 ColumnFamilyStore.java:411 - Initializing system_auth.roles INFO [InternalResponseStage:1] 2025-10-16 05:22:45,291 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_traces as storage service is not initialized INFO [InternalResponseStage:1] 2025-10-16 05:22:45,293 ColumnFamilyStore.java:411 - Initializing system_traces.events INFO [InternalResponseStage:1] 2025-10-16 05:22:45,297 ColumnFamilyStore.java:411 - Initializing system_traces.sessions INFO [MigrationStage:1] 2025-10-16 05:22:54,679 ViewManager.java:137 - Not submitting build tasks for views in keyspace reaper_db as storage service is not initialized INFO [MigrationStage:1] 2025-10-16 05:23:00,011 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration INFO [MigrationStage:1] 2025-10-16 05:23:00,271 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration_leader INFO [MigrationStage:1] 2025-10-16 05:23:01,390 ColumnFamilyStore.java:411 - Initializing reaper_db.running_reapers INFO [MigrationStage:1] 2025-10-16 05:23:02,001 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_unit_v1 INFO [MigrationStage:1] 2025-10-16 05:23:02,386 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_by_cluster_and_keyspace INFO [MigrationStage:1] 2025-10-16 05:23:02,932 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster INFO [MigrationStage:1] 2025-10-16 05:23:04,633 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_v1 INFO [MigrationStage:1] 2025-10-16 05:23:05,185 ColumnFamilyStore.java:411 - Initializing reaper_db.cluster INFO [InternalResponseStage:4] 2025-10-16 05:23:05,336 ColumnFamilyStore.java:411 - Initializing reaper_db.snapshot INFO [MigrationStage:1] 2025-10-16 05:23:05,727 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v1 INFO [MigrationStage:1] 2025-10-16 05:23:06,280 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run INFO [MigrationStage:1] 2025-10-16 05:23:06,395 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_unit INFO [MigrationStage:1] 2025-10-16 05:23:07,331 ColumnFamilyStore.java:411 - Initializing reaper_db.leader INFO [MigrationStage:1] 2025-10-16 05:23:10,936 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v2 INFO [MigrationStage:1] 2025-10-16 05:23:11,998 ColumnFamilyStore.java:411 - Initializing reaper_db.node_operations INFO [MigrationStage:1] 2025-10-16 05:23:13,464 ColumnFamilyStore.java:411 - Initializing reaper_db.diagnostic_event_subscription INFO [main] 2025-10-16 05:23:14,140 StorageService.java:1446 - JOINING: Starting to bootstrap... INFO [main] 2025-10-16 05:23:14,308 StreamResultFuture.java:90 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] Executing streaming plan for Bootstrap INFO [StreamConnectionEstablisher:1] 2025-10-16 05:23:14,310 StreamSession.java:266 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] Starting streaming to /10.0.0.241 INFO [StreamConnectionEstablisher:1] 2025-10-16 05:23:14,315 StreamCoordinator.java:264 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75, ID#0] Beginning stream session with /10.0.0.241 INFO [STREAM-IN-/10.0.0.241:7012] 2025-10-16 05:23:14,433 StreamResultFuture.java:173 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75 ID#0] Prepare completed. Receiving 2 files(2.450KiB), sending 0 files(0.000KiB) INFO [StreamConnectionEstablisher:2] 2025-10-16 05:23:14,436 StreamSession.java:266 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] Starting streaming to /10.0.0.38 INFO [StreamConnectionEstablisher:2] 2025-10-16 05:23:14,438 StreamCoordinator.java:264 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75, ID#0] Beginning stream session with /10.0.0.38 INFO [StreamReceiveTask:1] 2025-10-16 05:23:14,506 StreamResultFuture.java:187 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] Session with /10.0.0.241 is complete INFO [STREAM-IN-/10.0.0.38:7012] 2025-10-16 05:23:14,531 StreamResultFuture.java:187 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] Session with /10.0.0.38 is complete INFO [STREAM-IN-/10.0.0.38:7012] 2025-10-16 05:23:14,533 StreamResultFuture.java:219 - [Stream #36c437e0-aa50-11f0-b61b-4ff51e317e75] All sessions completed INFO [STREAM-IN-/10.0.0.38:7012] 2025-10-16 05:23:14,534 StorageService.java:1505 - Bootstrap completed! for the tokens [3865232099155310382, 5952851990590385243, 3240563202895670566, -1148421334932770250, -8225122346858646967, -2981274944881219572, -1122485615342174429, 4606393749740050827, 8476857092308191734, 3030588890644931510, -7219270434980484716, 5070364096661557868, 6466035134996645299, 7926987489270709243, -2859954181064986317, -4919629009555940254, -7640289567122276749, 8809828380506970146, 501522118087435457, 4384624772220997031, -1234469585154011053, -5601688050122438173, 1569218411336921228, 5237822838403033083, -4366652450095138937, 6480545411255989452, 233724486912859643, -4764034114964366773, 3695664036453928894, 8586368560238158908, 2892009789533283227, 7739613978207955714, -5526012803853879894, 4673046324942525286, 4180112473064902229, -184907261125272379, -675858733421903165, 4586218657462390316, 2769766058784149877, -1330952702924886258, -1715951035943847558, -546841866581489640, 1264796929087127800, -9076589835447936054, 4466311374397751910, -7678525806989887006, -2221599587513302893, 1673436004818470392, -3821730512769182254, 1525780980035280114, 5023295557663276336, 2836865183015497635, -3741224166047815792, -4234629247318450072, 779103728941929168, 1452742733987679947, 5847572623598330087, -1754398647267424933, 1023710574378442169, 3673474662376195433, 5543463606295835332, -392029869347299282, 4234162621617363577, -4813791488509643007, -3531636344889451511, 4936499777192508733, -7390836222643716542, 604780858942883370, -8385375980107047306, -6469095515862835779, 1957968757243793247, 3566269909373592856, -775307473100612482, 8585244511082622246, 7017783984413538767, -5120939234933750277, -597444285261257392, 6361256506469105905, 5235013197166584679, 2335198889157645778, 615842487493102065, -4629516826426624893, 3336554095973842127, -3293513670911660473, 2569633045331780282, -5999191512721443910, 77728647517695064, 8661073688073686817, -1873273496814937553, -5915571320365946764, -5815020260733370025, -7050290616687781989, 1221407080455191471, 8755057967890978299, -3611605159394968542, -2637700396918588570, 9033631800095651322, 7991106925625511207, 3960180770604945690, 7484953508503956574, 3157363434991878462, 3814921644584286801, 6910117400858066031, -5802498387193985971, -3824407340925453230, 5964198075320259850, 3246009993620776454, -514528772324163016, 6114558767938749951, 2882944036341699070, 2834965570769247106, -8345448320038572402, -2580223028720700975, -7905385344904519900, 7019506245055488793, 4181380819762484673, 216518570387412129, -295797731683704018, 7621886335007657630, -3653167035823285031, -4054503488734300494, -4536787329871394421, 1136683895312721610, 5026660684770498137, -7636523598898472952, -5703688183224259923, -5893416644393625715, 4702056683091557238, 712252187895585143, -7044172591420805961, 7971746203162708058, 6577035847637701233, -6963335994682042913, -2736979692793242142, -8918432740622949269, 1094968173268150266, 811453690600884590, -5726635861736788057, 2827774050137744428, 3275495825171781855, -4688858860399159876, -6676712882201264545, 7183478296953351798, -5747802695508707345, -7575920852765325395, -6398848602890001912, 995423515595922102, -402284296954540900, 7919135024846827983, 5773863082162895615, -4053252513129157340, 3991234581250826323, -1370510349670207173, -7003423919442890025, -2763372511037094326, 2546419350843339451, 1277973462797918947, 6815038078903172967, -8636126789088298431, 6322120371412670760, -655050326407568157, -3756604564210549745, -6293227713052197456, -190413793917570430, -2081873416720215450, 3528462649820529262, -684089748246519937, 7815233160955215235, -3236268020331366655, -2931450487814623647, 553397566181718291, 8683397936265121232, -873162731219207902, -3205885714676090598, 9030772085880197433, 855468030802882975, 7972973835592095602, 6455192313484128761, 6244809833295487994, -1566990106862673860, 6134443290042001001, -9209456500529342216, 5266390284938709008, -4803469450215782522, -7568208296246460626, 7327945051951642828, 6646375540639311784, 3103891783707885182, -6006848700026616754, 3746987514502335585, 3715361375216927128, 678099646916452557, 4095692096309391384, -1189642194835509329, 570885851432215747, 5494730972676586810, 4434161795058572617, -239058762841958178, -6091473887186623076, 3515279494042486721, 1431325337424230325, 154711859583290675, -6629033740889623217, 7892624676486468408, -6494573418728971940, 8959286372365191533, 2205869249560139165, 6015529335334559334, -8656051150893125007, -2976423085650191513, 6690376229882708260, 6434825285304077605, 5530481021300547055, 301852749886943130, -7305657642340971135, 3220767225740491228, 2968386505189148816, -4690695199001246429, -5249239732791944507, 7993199621543057712, -5747671335129439950, 3853143579874932842, 3964956702950941844, -8996780021335254741, 2217077795969506215, -6082264198933757746, 3752459118850308943, 2144710760865015885, 3664305369054715477, -3779375561136824309, -6964447335695006600, 6478721075999572362, 8227225588450768242, 1302080116720502536, 5029978552281406544, 4282086647115638124, -3745451879711745807, 604911619251529133, 6976327560632823693, 5110901750269140395, 6664447748659515253, -3151652423972229931, 7756793064548286700, 6832532424863132016, 4679568919244435889, 8423411312274512854, -45980178796501036, 4330917489463179661, 7624533946582132459, -7181361911088327991, -2224491597533857954, 3060562203777787315, 5340803614128372302, 3066023252551514619, -3080319591722426961, 5281792349889613330] INFO [main] 2025-10-16 05:23:14,537 StorageService.java:1446 - JOINING: Finish joining ring INFO [main] 2025-10-16 05:23:14,568 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='cluster') INFO [main] 2025-10-16 05:23:14,568 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='node_metrics_v1') INFO [main] 2025-10-16 05:23:14,568 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='snapshot') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_unit_v1') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_schedule_v1') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_run') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='diagnostic_event_subscription') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='leader') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='schema_migration') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_run_by_unit') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_schedule_by_cluster_and_keyspace') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='schema_migration_leader') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='running_reapers') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='node_operations') INFO [main] 2025-10-16 05:23:14,569 SecondaryIndexManager.java:509 - Executing pre-join post-bootstrap tasks for: CFS(Keyspace='reaper_db', ColumnFamily='repair_run_by_cluster') INFO [main] 2025-10-16 05:23:14,619 Gossiper.java:1692 - Waiting for gossip to settle... INFO [MigrationStage:1] 2025-10-16 05:23:14,970 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v3 INFO [MigrationStage:1] 2025-10-16 05:23:15,454 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster_v2 INFO [MigrationStage:1] 2025-10-16 05:23:17,037 ColumnFamilyStore.java:411 - Initializing reaper_db.running_repairs INFO [MigrationStage:1] 2025-10-16 05:23:17,501 ColumnFamilyStore.java:411 - Initializing reaper_db.percent_repaired_by_schedule INFO [main] 2025-10-16 05:23:22,620 Gossiper.java:1723 - No gossip backlog; proceeding INFO [main] 2025-10-16 05:23:22,880 NativeTransportService.java:70 - Netty using native Epoll event loop INFO [main] 2025-10-16 05:23:22,978 Server.java:155 - Using Netty Version: [netty-buffer=netty-buffer-4.1.39.Final.88c2a4c (repository: dirty), netty-codec=netty-codec-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-dns=netty-codec-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-haproxy=netty-codec-haproxy-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http=netty-codec-http-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http2=netty-codec-http2-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-memcache=netty-codec-memcache-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-mqtt=netty-codec-mqtt-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-redis=netty-codec-redis-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-smtp=netty-codec-smtp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-socks=netty-codec-socks-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-stomp=netty-codec-stomp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-xml=netty-codec-xml-4.1.39.Final.88c2a4c (repository: dirty), netty-common=netty-common-4.1.39.Final.88c2a4c (repository: dirty), netty-handler=netty-handler-4.1.39.Final.88c2a4c (repository: dirty), netty-handler-proxy=netty-handler-proxy-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver=netty-resolver-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver-dns=netty-resolver-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-tcnative=netty-tcnative-2.0.25.Final.c46c351, netty-transport=netty-transport-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-epoll=netty-transport-native-epoll-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-kqueue=netty-transport-native-kqueue-4.1.39.Final.88c2a4cab5 (repository: dirty), netty-transport-native-unix-common=netty-transport-native-unix-common-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-rxtx=netty-transport-rxtx-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-sctp=netty-transport-sctp-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-udt=netty-transport-udt-4.1.39.Final.88c2a4c (repository: dirty)] INFO [main] 2025-10-16 05:23:22,978 Server.java:156 - Starting listening for CQL clients on /10.0.0.242:9041 (unencrypted)... INFO [main] 2025-10-16 05:23:23,063 ThriftServer.java:116 - Binding thrift service to /10.0.0.242:9161 INFO [Thread-4] 2025-10-16 05:23:23,067 ThriftServer.java:133 - Listening for thrift clients... INFO [Native-Transport-Requests-1] 2025-10-16 05:23:27,144 MigrationManager.java:454 - Update table 'reaper_db/node_metrics_v3' From org.apache.cassandra.config.CFMetaData@737ac7b1[cfId=372da810-aa50-11f0-91c9-01e025d64bf5,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@30dc2bb[cfId=372da810-aa50-11f0-91c9-01e025d64bf5,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy, options={min_threshold=4, max_threshold=32, compaction_window_size=10, compaction_window_unit=MINUTES, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-10-16 05:23:42,128 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@2a287bac[cfId=47728b00-aa50-11f0-b61b-4ff51e317e75,ksName=config_db_uuid,cfName=obj_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:23:42,218 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_uuid_table INFO [Native-Transport-Requests-1] 2025-10-16 05:23:43,802 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@7a406fb[cfId=4871f9a0-aa50-11f0-b61b-4ff51e317e75,ksName=svc_monitor_keyspace,cfName=service_instance_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:23:43,897 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.service_instance_table INFO [MigrationStage:1] 2025-10-16 05:23:46,959 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_fq_name_table INFO [MigrationStage:1] 2025-10-16 05:23:48,361 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.pool_table INFO [Native-Transport-Requests-1] 2025-10-16 05:23:49,236 MigrationManager.java:454 - Update table 'config_db_uuid/obj_fq_name_table' From org.apache.cassandra.config.CFMetaData@5b3e858[cfId=4a3f66a0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3ff7f199[cfId=4a3f66a0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:23:50,279 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_shared_table INFO [Native-Transport-Requests-2] 2025-10-16 05:23:51,128 MigrationManager.java:454 - Update table 'svc_monitor_keyspace/pool_table' From org.apache.cassandra.config.CFMetaData@1cb9530[cfId=4b11f8e0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=svc_monitor_keyspace,cfName=pool_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@170f6137[cfId=4b11f8e0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=svc_monitor_keyspace,cfName=pool_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-10-16 05:23:52,135 MigrationManager.java:454 - Update table 'config_db_uuid/obj_shared_table' From org.apache.cassandra.config.CFMetaData@1d81635a[cfId=4c3d3270-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7abbd954[cfId=4c3d3270-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:23:53,227 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.loadbalancer_table INFO [HANDSHAKE-/10.0.0.242] 2025-10-16 05:23:55,180 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.242 INFO [MigrationStage:1] 2025-10-16 05:23:55,288 ColumnFamilyStore.java:411 - Initializing useragent.useragent_keyval_table INFO [Native-Transport-Requests-1] 2025-10-16 05:23:58,176 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@a3a0990[cfId=51034600-aa50-11f0-b61b-4ff51e317e75,ksName=svc_monitor_keyspace,cfName=healthmonitor_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:23:58,286 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.healthmonitor_table INFO [Native-Transport-Requests-1] 2025-10-16 05:24:00,965 MigrationManager.java:454 - Update table 'config_db_uuid/obj_fq_name_table' From org.apache.cassandra.config.CFMetaData@5b3e858[cfId=4a3f66a0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7b037492[cfId=4a3f66a0-aa50-11f0-9b3c-b3a6ba283ee6,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:24:05,424 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.route_target_table INFO [Native-Transport-Requests-1] 2025-10-16 05:24:06,062 MigrationManager.java:454 - Update table 'to_bgp_keyspace/route_target_table' From org.apache.cassandra.config.CFMetaData@3736c914[cfId=55431290-aa50-11f0-9b3c-b3a6ba283ee6,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@470de097[cfId=55431290-aa50-11f0-9b3c-b3a6ba283ee6,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:24:07,269 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_ip_address_table INFO [Native-Transport-Requests-1] 2025-10-16 05:24:09,153 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@7b18ec88[cfId=578e1400-aa50-11f0-b61b-4ff51e317e75,ksName=to_bgp_keyspace,cfName=service_chain_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:24:09,288 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_table INFO [MigrationStage:1] 2025-10-16 05:24:11,294 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_uuid_table INFO [Native-Transport-Requests-1] 2025-10-16 05:24:12,317 MigrationManager.java:454 - Update table 'to_bgp_keyspace/service_chain_uuid_table' From org.apache.cassandra.config.CFMetaData@713f9d6f[cfId=58c00450-aa50-11f0-9b3c-b3a6ba283ee6,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@21bf0dc[cfId=58c00450-aa50-11f0-9b3c-b3a6ba283ee6,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] + curl http://10.0.0.38:8071/webui/login.html % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1940 100 1940 0 0 8050 0 --:--:-- --:--:-- --:--:-- 8083 + export CASSANDRA_REAPER_JMX_KEY + run_service cassandra-reaper + [[ 10.0.0.38 == \1\0\.\0\.\0\.\2\4\2 ]] + echo 'Reaper started successfully' Reaper started successfully + [[ -n 1999 ]] + [[ -n 1999 ]] + local owner_opts=1999:1999 + mkdir -p /etc/contrail /var/lib/contrail + chown 1999:1999 /etc/contrail /var/lib/contrail + find /etc/contrail -uid 0 -exec chown 1999:1999 '{}' + + chmod 755 /etc/contrail + do_run_service cassandra-reaper + [[ -n 1999 ]] + [[ -n 1999 ]] + mkdir -p /var/crashes + chmod 777 /var/crashes ++ id -un 1999 + local user_name=contrail + export HOME=/home/contrail + HOME=/home/contrail + mkdir -p /home/contrail + chown -R 1999:1999 /home/contrail + exec setpriv --reuid 1999 --regid 1999 --clear-groups --no-new-privs cassandra-reaper Looking for reaper under /usr WARN [2025-10-16 05:24:36,874] [main] i.c.ReaperApplication - Reaper is ready to get things done! INFO [Native-Transport-Requests-1] 2025-10-16 05:25:16,040 MigrationManager.java:427 - Update Keyspace 'dm_keyspace' From KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-2] 2025-10-16 05:25:17,193 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@6b7499de[cfId=801c2880-aa50-11f0-b61b-4ff51e317e75,ksName=dm_keyspace,cfName=dm_pr_vn_ip_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:25:17,295 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_vn_ip_table INFO [MigrationStage:1] 2025-10-16 05:25:19,364 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_asn_table INFO [Native-Transport-Requests-1] 2025-10-16 05:25:20,198 MigrationManager.java:454 - Update table 'dm_keyspace/dm_pr_asn_table' From org.apache.cassandra.config.CFMetaData@644869fb[cfId=81571980-aa50-11f0-91c9-01e025d64bf5,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@64ac6198[cfId=81571980-aa50-11f0-91c9-01e025d64bf5,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:25:21,200 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_ni_ipv6_ll_table INFO [Native-Transport-Requests-1] 2025-10-16 05:25:23,146 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@7735c671[cfId=83a8aaa0-aa50-11f0-b61b-4ff51e317e75,ksName=dm_keyspace,cfName=dm_pnf_resource_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@d445ee2c, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-10-16 05:25:23,245 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pnf_resource_table INFO [Repair-Task-2] 2025-10-16 05:30:32,869 RepairRunnable.java:139 - Starting repair command #1 (3c44a550-aa51-11f0-b61b-4ff51e317e75), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 1, pull repair: false) INFO [Repair-Task-2] 2025-10-16 05:30:32,945 RepairSession.java:228 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] new session: will sync /10.0.0.242, /10.0.0.241, /10.0.0.38 on range [(-5677567552924720532,-5611418273250666821]] for reaper_db.[leader, repair_run_by_cluster, repair_run_by_cluster_v2, snapshot, repair_run, schema_migration_leader, cluster, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers, running_repairs, repair_schedule_v1, repair_run_by_unit, repair_unit_v1, diagnostic_event_subscription, percent_repaired_by_schedule] INFO [RepairJobTask:1] 2025-10-16 05:30:33,000 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:33,002 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,110 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,111 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,137 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,138 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,151 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:33,153 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:4] 2025-10-16 05:30:33,153 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-10-16 05:30:33,153 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:4] 2025-10-16 05:30:33,154 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] leader is fully synced INFO [RepairJobTask:4] 2025-10-16 05:30:33,164 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:4] 2025-10-16 05:30:33,164 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,166 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,166 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,171 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,171 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,174 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:33,175 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-10-16 05:30:33,175 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-10-16 05:30:33,175 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-10-16 05:30:33,175 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:33,260 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:33,260 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,263 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,263 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,265 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,265 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,269 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:33,270 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-10-16 05:30:33,270 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:4] 2025-10-16 05:30:33,270 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:3] 2025-10-16 05:30:33,270 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:33,280 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for snapshot (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:33,280 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,294 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,294 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,296 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,296 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,297 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:33,298 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:3] 2025-10-16 05:30:33,298 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:6] 2025-10-16 05:30:33,299 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:3] 2025-10-16 05:30:33,299 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] snapshot is fully synced INFO [RepairJobTask:6] 2025-10-16 05:30:33,406 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:6] 2025-10-16 05:30:33,406 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,410 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,410 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,412 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,412 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,414 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:33,415 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:3] 2025-10-16 05:30:33,415 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:5] 2025-10-16 05:30:33,415 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:3] 2025-10-16 05:30:33,415 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_run is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:33,419 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration_leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:33,419 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,427 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,427 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,429 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,429 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,431 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:33,432 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-10-16 05:30:33,432 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:7] 2025-10-16 05:30:33,435 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-10-16 05:30:33,435 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] schema_migration_leader is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:33,484 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:33,485 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,487 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,487 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,500 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,500 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,501 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:33,501 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:1] 2025-10-16 05:30:33,502 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:1] 2025-10-16 05:30:33,502 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:1] 2025-10-16 05:30:33,502 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] cluster is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:33,558 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:33,558 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,566 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,566 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,594 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,595 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,599 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.242 INFO [RepairJobTask:6] 2025-10-16 05:30:33,599 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-10-16 05:30:33,599 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-10-16 05:30:33,599 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-10-16 05:30:33,599 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:33,609 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:33,609 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,613 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,614 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,617 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,617 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,618 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:33,619 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:3] 2025-10-16 05:30:33,619 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:6] 2025-10-16 05:30:33,619 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:3] 2025-10-16 05:30:33,619 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] schema_migration is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:33,667 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_reapers (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:33,667 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,672 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,672 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,675 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,675 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,677 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:33,677 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:5] 2025-10-16 05:30:33,677 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:1] 2025-10-16 05:30:33,677 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:5] 2025-10-16 05:30:33,677 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] running_reapers is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:33,725 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_repairs (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:33,725 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,738 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,739 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,740 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,740 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,742 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:33,742 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:33,742 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:1] 2025-10-16 05:30:33,742 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:33,743 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] running_repairs is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:33,792 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:33,792 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,796 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,796 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,799 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,799 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,801 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.242 INFO [RepairJobTask:7] 2025-10-16 05:30:33,802 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-10-16 05:30:33,802 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:33,802 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-10-16 05:30:33,802 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_schedule_v1 is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:33,856 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:33,857 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,860 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,860 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,862 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,862 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,864 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:33,864 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:7] 2025-10-16 05:30:33,864 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-10-16 05:30:33,864 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:7] 2025-10-16 05:30:33,864 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_run_by_unit is fully synced INFO [RepairJobTask:7] 2025-10-16 05:30:33,923 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:7] 2025-10-16 05:30:33,924 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,932 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,933 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,936 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,937 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,941 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.242 INFO [RepairJobTask:7] 2025-10-16 05:30:33,941 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:33,941 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:1] 2025-10-16 05:30:33,941 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:33,941 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] repair_unit_v1 is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:33,945 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:33,945 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,947 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,947 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,950 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,950 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,953 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:33,953 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-10-16 05:30:33,953 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-10-16 05:30:33,953 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-10-16 05:30:33,953 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] diagnostic_event_subscription is fully synced INFO [RepairJobTask:7] 2025-10-16 05:30:33,958 RepairJob.java:234 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:7] 2025-10-16 05:30:33,958 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,968 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,968 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,971 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,971 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:33,973 RepairSession.java:180 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.242 INFO [RepairJobTask:7] 2025-10-16 05:30:33,973 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:1] 2025-10-16 05:30:33,973 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-10-16 05:30:33,973 SyncTask.java:66 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:2] 2025-10-16 05:30:33,973 RepairJob.java:143 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:33,974 RepairSession.java:270 - [repair #3c501700-aa51-11f0-b61b-4ff51e317e75] Session completed successfully INFO [RepairJobTask:2] 2025-10-16 05:30:33,975 RepairRunnable.java:261 - Repair session 3c501700-aa51-11f0-b61b-4ff51e317e75 for range [(-5677567552924720532,-5611418273250666821]] finished INFO [RepairJobTask:2] 2025-10-16 05:30:33,976 ActiveRepairService.java:452 - [repair #3c44a550-aa51-11f0-b61b-4ff51e317e75] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-10-16 05:30:33,985 RepairRunnable.java:343 - Repair command #1 finished in 1 second INFO [Repair-Task-3] 2025-10-16 05:30:37,554 RepairRunnable.java:139 - Starting repair command #2 (3f0f8520-aa51-11f0-b61b-4ff51e317e75), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 4, pull repair: false) INFO [Repair-Task-3] 2025-10-16 05:30:37,572 RepairSession.java:228 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] new session: will sync /10.0.0.242, /10.0.0.241, /10.0.0.38 on range [(-2498850951183766020,-2492512422536211886], (-1189642194835509329,-1172842514876860901], (-7003423919442890025,-6972365880096113925], (-1148421334932770250,-1128438368378221229]] for reaper_db.[leader, repair_run_by_cluster, repair_run_by_cluster_v2, snapshot, repair_run, schema_migration_leader, cluster, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers, running_repairs, repair_schedule_v1, repair_run_by_unit, repair_unit_v1, diagnostic_event_subscription, percent_repaired_by_schedule] INFO [RepairJobTask:2] 2025-10-16 05:30:37,641 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:37,641 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,647 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,648 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,651 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,651 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,654 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.242 INFO [RepairJobTask:1] 2025-10-16 05:30:37,663 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:1] 2025-10-16 05:30:37,663 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:37,663 RepairJob.java:257 - Validating /10.0.0.241 INFO [RepairJobTask:4] 2025-10-16 05:30:37,663 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-10-16 05:30:37,663 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:4] 2025-10-16 05:30:37,663 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] leader is fully synced INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,667 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,667 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,670 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,670 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,672 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:37,674 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-10-16 05:30:37,675 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:3] 2025-10-16 05:30:37,675 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-10-16 05:30:37,675 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:37,684 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:37,684 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,691 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,691 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,695 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,695 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,699 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:37,700 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:1] 2025-10-16 05:30:37,700 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-10-16 05:30:37,700 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:1] 2025-10-16 05:30:37,700 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:37,707 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for snapshot (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:37,707 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,717 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,718 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,720 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,720 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,721 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:37,722 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:6] 2025-10-16 05:30:37,723 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:5] 2025-10-16 05:30:37,723 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:6] 2025-10-16 05:30:37,723 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] snapshot is fully synced INFO [RepairJobTask:6] 2025-10-16 05:30:37,829 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:6] 2025-10-16 05:30:37,829 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,843 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,843 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,845 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,846 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,848 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:37,849 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:3] 2025-10-16 05:30:37,849 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:7] 2025-10-16 05:30:37,850 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:5] 2025-10-16 05:30:37,850 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_run is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:37,861 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration_leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:37,861 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,865 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,866 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,869 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,869 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,870 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:37,870 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:6] 2025-10-16 05:30:37,870 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-10-16 05:30:37,870 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-10-16 05:30:37,871 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] schema_migration_leader is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:37,876 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:37,876 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,879 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,879 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,882 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,882 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,883 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:37,884 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:2] 2025-10-16 05:30:37,884 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:3] 2025-10-16 05:30:37,884 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:1] 2025-10-16 05:30:37,884 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] cluster is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:37,892 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:37,892 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,897 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,898 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,904 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,904 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,907 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:37,908 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:1] 2025-10-16 05:30:37,908 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-10-16 05:30:37,908 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:2] 2025-10-16 05:30:37,908 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:37,914 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:37,914 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,927 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,927 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,933 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,933 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,935 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:37,937 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:5] 2025-10-16 05:30:37,937 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:1] 2025-10-16 05:30:37,937 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:5] 2025-10-16 05:30:37,938 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] schema_migration is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:37,946 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_reapers (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:37,946 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,966 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,966 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,973 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,974 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:37,975 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:37,975 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:2] 2025-10-16 05:30:37,976 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:3] 2025-10-16 05:30:37,976 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:2] 2025-10-16 05:30:37,976 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] running_reapers is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:38,031 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_repairs (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:38,032 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,044 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,044 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,050 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,050 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,053 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.242 INFO [RepairJobTask:1] 2025-10-16 05:30:38,053 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:38,053 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:3] 2025-10-16 05:30:38,053 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:38,053 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] running_repairs is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:38,067 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:38,067 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,072 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,073 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,077 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,078 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,081 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.242 INFO [RepairJobTask:1] 2025-10-16 05:30:38,081 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:38,081 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:3] 2025-10-16 05:30:38,081 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:38,081 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_schedule_v1 is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:38,086 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:38,086 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,089 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,089 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,093 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,093 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,094 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:38,095 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-10-16 05:30:38,095 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-10-16 05:30:38,095 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:3] 2025-10-16 05:30:38,095 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_run_by_unit is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:38,107 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:38,107 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,110 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,110 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,114 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,115 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,117 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:38,117 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:2] 2025-10-16 05:30:38,117 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-10-16 05:30:38,117 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-10-16 05:30:38,117 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] repair_unit_v1 is fully synced INFO [RepairJobTask:6] 2025-10-16 05:30:38,134 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:6] 2025-10-16 05:30:38,134 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,137 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,137 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,141 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,141 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,143 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:38,143 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:4] 2025-10-16 05:30:38,143 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-10-16 05:30:38,143 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:4] 2025-10-16 05:30:38,144 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] diagnostic_event_subscription is fully synced INFO [RepairJobTask:4] 2025-10-16 05:30:38,150 RepairJob.java:234 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:4] 2025-10-16 05:30:38,151 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,153 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,153 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,157 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,157 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:38,159 RepairSession.java:180 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:38,159 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:1] 2025-10-16 05:30:38,159 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-10-16 05:30:38,159 SyncTask.java:66 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:1] 2025-10-16 05:30:38,159 RepairJob.java:143 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:38,161 RepairSession.java:270 - [repair #3f124440-aa51-11f0-b61b-4ff51e317e75] Session completed successfully INFO [RepairJobTask:1] 2025-10-16 05:30:38,161 RepairRunnable.java:261 - Repair session 3f124440-aa51-11f0-b61b-4ff51e317e75 for range [(-2498850951183766020,-2492512422536211886], (-1189642194835509329,-1172842514876860901], (-7003423919442890025,-6972365880096113925], (-1148421334932770250,-1128438368378221229]] finished INFO [RepairJobTask:1] 2025-10-16 05:30:38,162 ActiveRepairService.java:452 - [repair #3f0f8520-aa51-11f0-b61b-4ff51e317e75] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-10-16 05:30:38,166 RepairRunnable.java:343 - Repair command #2 finished in 0 seconds INFO [Repair-Task-4] 2025-10-16 05:30:42,992 RepairRunnable.java:139 - Starting repair command #3 (424d4b00-aa51-11f0-b61b-4ff51e317e75), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 3, pull repair: false) INFO [Repair-Task-4] 2025-10-16 05:30:43,007 RepairSession.java:228 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] new session: will sync /10.0.0.242, /10.0.0.241, /10.0.0.38 on range [(6679593360763368620,6690376229882708260], (1221407080455191471,1264796929087127800], (-6440370339942402911,-6428124747150390857]] for reaper_db.[leader, repair_run_by_cluster, repair_run_by_cluster_v2, snapshot, repair_run, schema_migration_leader, cluster, repair_schedule_by_cluster_and_keyspace, schema_migration, running_reapers, running_repairs, repair_schedule_v1, repair_run_by_unit, repair_unit_v1, diagnostic_event_subscription, percent_repaired_by_schedule] INFO [RepairJobTask:3] 2025-10-16 05:30:43,044 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,044 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,047 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,047 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,069 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,069 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,073 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for leader from /10.0.0.242 INFO [RepairJobTask:1] 2025-10-16 05:30:43,075 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:3] 2025-10-16 05:30:43,076 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for leader INFO [RepairJobTask:5] 2025-10-16 05:30:43,076 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-10-16 05:30:43,076 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] leader is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:43,087 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,087 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,093 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,093 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,097 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,097 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,099 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster from /10.0.0.242 INFO [RepairJobTask:1] 2025-10-16 05:30:43,100 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-10-16 05:30:43,100 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-10-16 05:30:43,100 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster INFO [RepairJobTask:3] 2025-10-16 05:30:43,107 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:43,113 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,113 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,115 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,115 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,158 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,158 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,161 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:43,163 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-10-16 05:30:43,163 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-10-16 05:30:43,163 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:2] 2025-10-16 05:30:43,163 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:43,168 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for snapshot (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:43,171 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,182 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,182 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,191 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,191 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,193 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for snapshot from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:43,196 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:5] 2025-10-16 05:30:43,196 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:2] 2025-10-16 05:30:43,197 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for snapshot INFO [RepairJobTask:2] 2025-10-16 05:30:43,197 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] snapshot is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:43,258 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,258 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,264 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,264 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,315 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,315 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,316 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:43,317 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:5] 2025-10-16 05:30:43,318 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:5] 2025-10-16 05:30:43,318 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run INFO [RepairJobTask:5] 2025-10-16 05:30:43,322 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_run is fully synced INFO [RepairJobTask:4] 2025-10-16 05:30:43,333 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration_leader (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:4] 2025-10-16 05:30:43,333 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,336 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,336 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,341 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,342 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,345 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration_leader from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:43,345 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-10-16 05:30:43,345 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-10-16 05:30:43,345 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-10-16 05:30:43,345 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] schema_migration_leader is fully synced INFO [RepairJobTask:4] 2025-10-16 05:30:43,351 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for cluster (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:4] 2025-10-16 05:30:43,352 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,354 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,354 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,360 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,361 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,363 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for cluster from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:43,364 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:2] 2025-10-16 05:30:43,365 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:2] 2025-10-16 05:30:43,365 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for cluster INFO [RepairJobTask:2] 2025-10-16 05:30:43,365 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] cluster is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:43,377 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:43,378 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,381 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,381 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,386 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,386 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,387 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:43,387 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-10-16 05:30:43,388 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:5] 2025-10-16 05:30:43,388 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:3] 2025-10-16 05:30:43,388 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:43,398 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for schema_migration (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,399 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,403 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,403 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,406 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,406 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,407 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for schema_migration from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:43,408 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:3] 2025-10-16 05:30:43,408 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for schema_migration INFO [RepairJobTask:2] 2025-10-16 05:30:43,408 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:3] 2025-10-16 05:30:43,409 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] schema_migration is fully synced INFO [RepairJobTask:3] 2025-10-16 05:30:43,411 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_reapers (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:3] 2025-10-16 05:30:43,411 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,449 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,449 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,457 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,459 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,460 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_reapers from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:43,461 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:1] 2025-10-16 05:30:43,461 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_reapers INFO [RepairJobTask:4] 2025-10-16 05:30:43,461 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:1] 2025-10-16 05:30:43,461 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] running_reapers is fully synced INFO [RepairJobTask:1] 2025-10-16 05:30:43,516 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for running_repairs (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:1] 2025-10-16 05:30:43,516 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,522 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,522 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,526 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,526 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,529 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for running_repairs from /10.0.0.242 INFO [RepairJobTask:4] 2025-10-16 05:30:43,531 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:43,531 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:3] 2025-10-16 05:30:43,531 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for running_repairs INFO [RepairJobTask:5] 2025-10-16 05:30:43,531 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] running_repairs is fully synced INFO [RepairJobTask:5] 2025-10-16 05:30:43,541 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:5] 2025-10-16 05:30:43,541 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,563 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,563 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,568 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,568 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,570 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_schedule_v1 from /10.0.0.242 INFO [RepairJobTask:2] 2025-10-16 05:30:43,570 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-10-16 05:30:43,570 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:6] 2025-10-16 05:30:43,570 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-10-16 05:30:43,570 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_schedule_v1 is fully synced INFO [RepairJobTask:2] 2025-10-16 05:30:43,599 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:2] 2025-10-16 05:30:43,599 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,603 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,603 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,607 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,607 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,608 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_run_by_unit from /10.0.0.242 INFO [RepairJobTask:3] 2025-10-16 05:30:43,609 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-10-16 05:30:43,609 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:7] 2025-10-16 05:30:43,610 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_run_by_unit INFO [RepairJobTask:6] 2025-10-16 05:30:43,610 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_run_by_unit is fully synced INFO [RepairJobTask:6] 2025-10-16 05:30:43,619 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:6] 2025-10-16 05:30:43,619 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,629 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,629 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,635 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,635 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,637 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for repair_unit_v1 from /10.0.0.242 INFO [RepairJobTask:5] 2025-10-16 05:30:43,637 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:43,637 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:43,637 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-10-16 05:30:43,638 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] repair_unit_v1 is fully synced INFO [RepairJobTask:7] 2025-10-16 05:30:43,642 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:7] 2025-10-16 05:30:43,643 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,646 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,646 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,667 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,667 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,668 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for diagnostic_event_subscription from /10.0.0.242 INFO [RepairJobTask:7] 2025-10-16 05:30:43,668 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:4] 2025-10-16 05:30:43,668 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:6] 2025-10-16 05:30:43,669 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-10-16 05:30:43,669 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] diagnostic_event_subscription is fully synced INFO [RepairJobTask:7] 2025-10-16 05:30:43,674 RepairJob.java:234 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.241, /10.0.0.38, /10.0.0.242]) INFO [RepairJobTask:7] 2025-10-16 05:30:43,674 RepairJob.java:257 - Validating /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,678 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.241 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,678 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,706 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,706 RepairJob.java:270 - Validating /10.0.0.242 INFO [AntiEntropyStage:1] 2025-10-16 05:30:43,709 RepairSession.java:180 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Received merkle tree for percent_repaired_by_schedule from /10.0.0.242 INFO [RepairJobTask:7] 2025-10-16 05:30:43,709 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-10-16 05:30:43,709 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.241 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-10-16 05:30:43,709 SyncTask.java:66 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Endpoints /10.0.0.38 and /10.0.0.242 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-10-16 05:30:43,709 RepairJob.java:143 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:4] 2025-10-16 05:30:43,710 RepairSession.java:270 - [repair #424f94f0-aa51-11f0-b61b-4ff51e317e75] Session completed successfully INFO [RepairJobTask:4] 2025-10-16 05:30:43,710 RepairRunnable.java:261 - Repair session 424f94f0-aa51-11f0-b61b-4ff51e317e75 for range [(6679593360763368620,6690376229882708260], (1221407080455191471,1264796929087127800], (-6440370339942402911,-6428124747150390857]] finished INFO [RepairJobTask:4] 2025-10-16 05:30:43,711 ActiveRepairService.java:452 - [repair #424d4b00-aa51-11f0-b61b-4ff51e317e75] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-10-16 05:30:43,715 RepairRunnable.java:343 - Repair command #3 finished in 0 seconds INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,753 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,782 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,809 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,828 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,847 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,872 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,887 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,913 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,930 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:47,996 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,013 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,040 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,064 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,135 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,248 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,292 Validator.java:281 - [repair #4519d880-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:48,319 ActiveRepairService.java:452 - [repair #45174070-aa51-11f0-9b3c-b3a6ba283ee6] Not a global repair, will not do anticompaction INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,284 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,304 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,321 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,344 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,356 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,366 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,378 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,437 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,449 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,521 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,557 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,588 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,608 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,620 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,637 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,654 Validator.java:281 - [repair #485cce80-aa51-11f0-91c9-01e025d64bf5] Sending completed merkle tree to /10.0.0.38 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-10-16 05:30:53,668 ActiveRepairService.java:452 - [repair #48552d60-aa51-11f0-91c9-01e025d64bf5] Not a global repair, will not do anticompaction INFO [AntiEntropyStage:1] 2025-10-16 05:30:57,967 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:57,985 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,006 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,054 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,065 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,086 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,106 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,134 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,156 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,227 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,267 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,314 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,343 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,358 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,425 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,450 Validator.java:281 - [repair #4b2bcd00-aa51-11f0-9b3c-b3a6ba283ee6] Sending completed merkle tree to /10.0.0.241 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-10-16 05:30:58,468 ActiveRepairService.java:452 - [repair #4b295c00-aa51-11f0-9b3c-b3a6ba283ee6] Not a global repair, will not do anticompaction