multicloud365
  • Home
  • Cloud Architecture
    • OCI
    • GCP
    • Azure
    • AWS
    • IAC
    • Cloud Networking
    • Cloud Trends and Innovations
    • Cloud Security
    • Cloud Platforms
  • Data Management
  • DevOps and Automation
    • Tutorials and How-Tos
  • Case Studies and Industry Insights
    • AI and Machine Learning in the Cloud
No Result
View All Result
  • Home
  • Cloud Architecture
    • OCI
    • GCP
    • Azure
    • AWS
    • IAC
    • Cloud Networking
    • Cloud Trends and Innovations
    • Cloud Security
    • Cloud Platforms
  • Data Management
  • DevOps and Automation
    • Tutorials and How-Tos
  • Case Studies and Industry Insights
    • AI and Machine Learning in the Cloud
No Result
View All Result
multicloud365
No Result
View All Result

Managed Service for Kafka benchmarking and scaling steering

admin by admin
June 15, 2025
in GCP
0
Introducing Google Cloud Serverless for Apache Spark in BigQuery
399
SHARES
2.3k
VIEWS
Share on FacebookShare on Twitter


Necessary concerns

To attain optimum client throughput, the fetch-size property is essential for tuning. The default fetch-size configuration is basically decided by your consumption and throughput wants, and may vary from as much as 1MB for smaller messages to 1-50MB for bigger ones. It is advisable to research the consequences of various fetch sizes on each utility responsiveness and throughput. By rigorously documenting these checks and analyzing the ensuing data, you’ll be able to pinpoint efficiency limitations and refine your settings accordingly.

The best way to benchmark throughput and latencies

Benchmarking the producer

When conducting checks to measure the throughput and latencies of Kafka producers, the important thing parameters are batch.measurement, or the utmost measurement of a batch of messages, and linger.ms, the utmost time to attend for a batch to fill earlier than sending. For the needs of this benchmark, we advise retaining acks at 1 (acknowledgment from the chief dealer) to stability sturdiness and efficiency. This helps us to estimate the anticipated throughput and latencies for a producer. Be aware that message measurement is stored fixed as 1KB.

Throughput (messages/s)

Throughput (MBs)

Latency(ms)

ack(=1)

batch_size

linger_ms

48049

45

608

Chief

1KB

10

160694

153

171

Chief

10KB

10

117187

111

268

Chief

100KB

10

111524

106

283

Chief

100KB

100

Evaluation and findings

  • The impression of batch measurement: As anticipated, rising batch measurement typically results in greater throughput (messages/s and MBs). We see a major soar in throughput as we transfer from 1KB to 10KB batch sizes. Nonetheless, additional rising the batch measurement to 100KB doesn’t present a major enchancment in throughput. This means that an optimum batch measurement exists past which additional will increase might not yield substantial throughput positive aspects.

  • Impression of linger time: Rising the linger time from 10ms to 100ms with a 100KB batch measurement barely lowered throughput (from 117,187 to 111,524 messages/s). This means that, on this state of affairs, an extended linger time may not be a lot helpful for maximizing throughput.

  • Latency concerns: Latency tends to extend with bigger batch sizes. It’s because messages wait longer to be included in a bigger batch earlier than being despatched. That is clearly seen when batch_size is elevated from 10KB to 100KB.

Collectively, these findings spotlight the significance of cautious tuning when configuring Kafka producers. Discovering the optimum stability between batch.measurement and linger.ms is essential for reaching desired throughput and latency targets.

Benchmarking for client

To evaluate client efficiency, we carried out a sequence of experiments utilizing kafka-consumer-perf-test, systematically various the fetch measurement.

Throughput(messages/sec)

Throughput(MBs)

fetch-size

2825

2.6951

10KB

3645

3.477

100KB

18086

17.8

1MB

49048

46

10MB

61334

58

100MB

62562

60

500MB

Evaluation and findings

  • Impression of fetch measurement on throughput: The outcomes clearly reveal a powerful correlation between fetch.measurement and client throughput. As we enhance the fetch measurement, each message throughput (messages/s) and information throughput (MBs) enhance considerably. It’s because bigger fetch sizes enable the buyer to retrieve extra messages in a single request, decreasing the overhead of frequent requests and enhancing information switch effectivity.

  • Diminishing returns: Whereas rising fetch.measurement typically improves throughput, we observe diminishing returns as we transfer past 100MB. The distinction in throughput between 100MB and 500MB is just not vital, suggesting that there is a level the place additional rising the fetch measurement offers minimal extra profit.

Scaling the Google Managed Service for Apache Kafka

Based mostly on some extra experiments, we explored optimum configurations for the managed Kafka cluster. Please be aware that for this train, we stored message measurement as 1KB, batch measurement as 10KB, the subject has 1000 partitions, and the replication quantity is 3. The outcomes have been as follows.

Producer threads

cluster_bytes_in_count (MBs)

CPU Util

Reminiscence Util

vCPU

Reminiscence

1

56

98%

58%

3

12gb

1

61

24%

41%

12

48gb

2

104

56%

57%

12

48gb

4

199

64%

60%

12

48gb

Scaling your managed Kafka cluster successfully is essential to make sure optimum efficiency as your necessities develop. To find out the proper cluster configuration, we carried out experiments with various numbers of producer threads, vCPUs, and reminiscence. Our findings point out that vertical scaling, by rising vCPUs and reminiscence from 3 vCPUs/12GB to 12 vCPUs/48GB, considerably improved useful resource utilization. With two producer threads, the cluster’s byte_in_count metric doubled and CPU utilization elevated to 56% from 24%. Your throughput necessities play a significant position. With 12 vCPUs/48GB, transferring from 2 to 4 producer threads almost doubled the cluster’s bytes_in_count. You additionally want to observe useful resource utilization to keep away from bottlenecks, as rising throughput can enhance CPU and reminiscence utilization. In the end, optimizing managed Kafka service efficiency requires a cautious stability between vertical scaling of the cluster and your throughput necessities, tailor-made to your particular workload and useful resource constraints.

Construct the Kafka cluster you want

In conclusion, optimizing your Google Cloud Managed Service for Apache Kafka deployment includes an intensive understanding of producer and client conduct, cautious benchmarking, and strategic scaling. By actively monitoring useful resource utilization and adjusting your configurations based mostly in your particular workload calls for, you’ll be able to guarantee your managed Kafka clusters ship the excessive throughput and low latency required on your real-time information streaming purposes.

Inquisitive about diving deeper? Discover the sources and documentations linked under:

Tags: benchmarkingGuidanceKafkaManagedScalingService
Previous Post

#AI horizons 25-05 – Coding within the Age of Brokers: How Windsurf, GitHub, and Claude Are Reprogramming Software program Improvement

Next Post

20 Free Spreadsheet Templates for Excel, Google Sheets & Stackby

Next Post
20 Free Spreadsheet Templates for Excel, Google Sheets & Stackby

20 Free Spreadsheet Templates for Excel, Google Sheets & Stackby

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Trending

Main CASB distributors | ManagedMethods

Main CASB distributors | ManagedMethods

March 23, 2025
B2B Analysis in a G-Zero World: Challenges & Options

B2B Analysis in a G-Zero World: Challenges & Options

April 7, 2025
What’s SIEM? Safety Info and Occasion Administration Defined

Prime Purposes of Generative AI Remodeling Industries

May 9, 2025
Plans, Critiques and Hidden Prices

Plans, Critiques and Hidden Prices

April 25, 2025
Easy methods to Choose the Proper Multi-Cloud Atmosphere for Your Enterprise

Easy methods to Choose the Proper Multi-Cloud Atmosphere for Your Enterprise

May 24, 2025
Return of The Frugal Architect(s)

Return of The Frugal Architect(s)

January 25, 2025

MultiCloud365

Welcome to MultiCloud365 — your go-to resource for all things cloud! Our mission is to empower IT professionals, developers, and businesses with the knowledge and tools to navigate the ever-evolving landscape of cloud technology.

Category

  • AI and Machine Learning in the Cloud
  • AWS
  • Azure
  • Case Studies and Industry Insights
  • Cloud Architecture
  • Cloud Networking
  • Cloud Platforms
  • Cloud Security
  • Cloud Trends and Innovations
  • Data Management
  • DevOps and Automation
  • GCP
  • IAC
  • OCI

Recent News

What The Knowledge Actually Says

What The Knowledge Actually Says

July 19, 2025
Construct real-time journey suggestions utilizing AI brokers on Amazon Bedrock

Construct real-time journey suggestions utilizing AI brokers on Amazon Bedrock

July 19, 2025
  • About Us
  • Privacy Policy
  • Disclaimer
  • Contact

© 2025- https://multicloud365.com/ - All Rights Reserved

No Result
View All Result
  • Home
  • Cloud Architecture
    • OCI
    • GCP
    • Azure
    • AWS
    • IAC
    • Cloud Networking
    • Cloud Trends and Innovations
    • Cloud Security
    • Cloud Platforms
  • Data Management
  • DevOps and Automation
    • Tutorials and How-Tos
  • Case Studies and Industry Insights
    • AI and Machine Learning in the Cloud

© 2025- https://multicloud365.com/ - All Rights Reserved