A little AWS EC2 benchmarking of drives #6181
jhughes2112
started this conversation in
Show and tell
Replies: 2 comments
-
One issue is that if you are saturating the CPU or IO, Redpanda doesn't handle this well: #608. You are hitting this case in the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I spent most of the day trying out various things on EC2. Since data is time-consuming and hard to come by, but fun to absorb quickly, I'll share what I discovered. Running on an i4i.2xlarge with 8 CPUs, I used 6 for RedPanda and used 20GB of memory, 1 topic with 20 partitions. I re-ran
rpk iotune
each time I changed the data folder and checked the drive speed withdd
as well. Furthermore, I compared Ubuntu 22.04LTS against Amazon's Linux AMI based on Debian. The numbers turned out identical, so pick whichever distro you like. I ran RedPanda like this:sudo /usr/bin/rpk redpanda start --smp 6 --memory 20G --reserve-memory 0M --node-id 1 --kafka-addr bob://0.0.0.0:9092 --advertise-kafka-addr bob://x.x.x.x:9092
and created the topic as such:sudo rpk topic create mytopic -p 20
I tested production of messages with the following:sudo docker run -it --rm salaholabi/rdkafka_performance ./rdkafka_performance -P -b x.x.x.x:9092 -t mytopic -s 275 -c 10000000 -a -1 -u
Looks like direct-mounted NVMe is the way to go, followed closely (but expensively) with IO2, and unacceptably gp3 lags behind. Seems like the gp3 drive was fast at first but got slower after a while... not sure what happened there. Anyway, I hope this helps someone.
Beta Was this translation helpful? Give feedback.
All reactions