I got this error running an Spark process with the spark-cassandra-connector. I have the following configuration:
spark.cassandra.output.consistency.level = "LOCAL_ONE"
spark.cassandra.output.concurrent.writes = "1"
spark.cassandra.output.batch.grouping.buffer.size = "10"
spark.cassandra.output.batch.size.rows = "10000"
spark.cassandra.output.batch.grouping.key = "replica_set"
spark.cassandra.output.throughput_mb_per_sec = "10"
I believe this might have been cause because the size of output.batch.size.rows = "10000" exceeded the length 268435456 which is defined by parameter "native_transport_max_frame_size_in_mb: 256". Would this be avoided by setting the output.batch.size.bytes to certain amount of bytes instead of rows? Therefore my question would be what is the ideal value to have high throughput?
At the time of the error here is my write request per minute:
Thanks