Bringing together the Apache Cassandra experts from the community and DataStax.

Want to learn? Have a question? Want to share your expertise? You are in the right place!

Not sure where to begin? Getting Started

 

question

gmldba_107428 avatar image
gmldba_107428 asked gmldba_107428 commented

Shouldn't Cassandra eliminate duplicates in compaction? Part 2

I know already asked this question in https://community.datastax.com/questions/10040/duplicate-error.html?childToView=10101.
But still something bothers me, my cluster has been 3.11 since more than 2 years. Yet, I still see this error:

WARN [Native-Transport-Requests-2] 2021-01-15 12:38:43,830 DuplicateRowChecker.java:96 - Detected 1 duplicate rows for release-metric-d2c23ed5-e281-49cf-9970-35e842b9d82e:data_points:2019-01-17 00\:00Z during Read.

I perform weekly full repairs and my queries are consistency level quorum. Why am I seeing this warning?

compaction
10 |1000 characters needed characters left characters exceeded

Up to 8 attachments (including images) can be used with a maximum of 1.0 MiB each and 10.0 MiB total.

1 Answer

Erick Ramirez avatar image
Erick Ramirez answered gmldba_107428 commented

As I explained in your other post (#10040), it is a warning message and not an error.

Cassandra has detected duplicate rows in partition with key release-metric-d2c23ed5-e281-49cf-9970-35e842b9d82e:data_points:2019-01-17 00:00Z. This means that the fix which prevents duplicate rows from getting returned is working. Cheers!

1 comment Share
10 |1000 characters needed characters left characters exceeded

Up to 8 attachments (including images) can be used with a maximum of 1.0 MiB each and 10.0 MiB total.

Does this means that Cassandra is not sending old data or duplicates?

0 Likes 0 ·