Query Nodes Failed to Balance Data Among all Nodes Evenly #38114
Replies: 3 comments 1 reply
-
There are only 2 collections, each collection has 10 partitions. Each partition has around 100 segments. |
Beta Was this translation helpful? Give feedback.
-
Updates: because the query node crashed, and it can't emit insight to the coordinator, causing the failure to redistribute the workloads. |
Beta Was this translation helpful? Give feedback.
-
@xiaobingxia-at Can I ask your ingestion/deletion speed after scaling? |
Beta Was this translation helpful? Give feedback.
-
Hello, I'm using HNSW/IVF-SQ8 + MMAP + nvme local disk. I was ingesting data with 1 query node, and then later on, I stopped the ingestion, and scaled the query nodes from 1 to 4. Then I restarted the ingesting data process, but this process will delete the existing data incrementally, and then re-ingest the data with same id. Then I see that the very original query node's memory usage gradually grows to 100GB, while others query nodes are staying at the 5GB. And finally, the ingestion reported error that we can't ingest data any more.
Does any one have any insight why one query node becomes really hot and it didn't distribute the data ingestion workload to other query nodes?
Related Config:
Beta Was this translation helpful? Give feedback.
All reactions