This document describes at a high level how producing and consuming works, and links to pkg.go.dev documentation for you to understand more on the mentioned functions/methods and how they can be used.
Code for both producing and consuming can be seen in the examples directory, particularly in the transaction examples.
The client provides three methods to produce, Produce
,
ProduceSync
, and TryProduce
. The first allows for
asynchronous production, the second for synchronous, and the third for async
while also failing a record immediately if the maximum records
are buffered. These methods are also available on GroupTransactSession
if you are using that for EOS.
Everything is produced through a Record
. You can produce to multiple
topics by creating a record in full (i.e., with a Value
or Key
or
Headers
) and then setting the Topic
field, but if you are only ever
producing to one topic, you can use the client's DefaultProduceTopic
option.
You can still use this option even when producing to multiple topics; the
option only applies to records that have an empty topic.
There exist a few small helpers to create records out of slices or strings:
The string functions should only be used if you do not touch the Key
and
Value
fields after the record is created, because the string functions use
the unsafe
package to convert the strings to slices without allocating.
Lastly, if you are producing asynchronously in batches and only want to know
whether the batch errored at all, there exists a FirstErrPromise
type to
help eliminate promise boilerplate.
By default, kgo uses idempotent production. This can be disabled with the
DisableIdempotentWrite
option, but this should really only be necessary
if you want to produce with no ack required or with only leader acks required.
The default is to always retry records forever, but this can be dropped with
the RecordRetries
and RecordDeliveryTimeout
options, as well as with
the context that you use for producing a record. A record will only be aborted
if it is safe to do so without messing up the client's sequence numbers. Thus,
a record can only be aborted if it has never been produced or if it knows that
it received a successful response from its last produce attempt (even if that
response indicated an error on that partition). If a record is ever failed, all
records buffered on the same partition are failed.
As mentioned above, kgo supports EOS. Because there are a lot of corner cases
around transactions, this client favors a "if we maybe should abort, abort"
approach. This client provides a GroupTransactSession
type that is used
to manage consume-modify-produce transactions. Any time it is possible that
the transaction should be aborted, the session sets its internal abort state.
This may mean you will end up re-processing records more than necessary, but
in general this should only happen on group rebalances, which should be rare.
Producer-only transactions are also supported. This is just a simple extension of the idempotent producer except with a manual begin transaction and end transaction call whenever appropriate.
To see more documentation about transactions and EOS, see the transactions page.
Producer latency can be modified by adding a linger. By default, there is no linger and records are sent as soon as they are published. In a high throughput scenario, this is fine and will not lead to single-record batches, but in a low throughput scenario it may be worth it to add lingering.
As well, it is possible to completely disable auto-flushing and instead only
have manual flushes with the
ManualFlushing
option. This allows you to buffer as much as you want before flushing in one
go. However, with this option, you likely want to consider the
MaxBufferedRecords
option.
franz-go supports consuming partitions directly, consuming as a part of a consumer group, and consuming in a group for EOS. Consuming can also be done via regex to match certain topics to consume.
To consume partitions directly, use ConsumeTopics
or ConsumePartitions
. Otherwise, use
ConsumeTopics
with ConsumerGroup
for group consuming or NewGroupTransactSession
for group consuming for EOS.
The default consumer group balancer is the new "cooperative-sticky" balancer. This is not compatible with historical balancers (sticky, range, roundrobin). If you wish to use this client with another client that uses a historical balancer, you must set the balancers option.
By default, the group consumer will autocommit every 5s, commit whenever a
rebalance happens (in OnPartitionsRevoked
), and will issue a blocking
commit when leaving the group. For most purposes, this can suffice. The default
commit logs any errors encountered, but this can be overridden with the
AutoCommitCallback
option or by disabling autocommit and instead committing
yourself.
Unlike Sarama or really most Kafka clients, this client manages the consumer
group completely independently from consuming itself. More to the point, a
revoke can happen at any time and if you need to stop consuming or do some
cleanup on a revoke, you must set a callback that will not return until you
are ready for the group to be rejoined. Even more to the point, if you are
manually committing offsets, you must commit in your OnPartitionsRevoked
,
or you must abandon your work after the revoke finishes, because otherwise you
may be working on partitions that moved to another client. When you are done
consuming, before you shut down, you must perform a blocking commit. If you
rely on the default options and do not commit yourself, all of this is
automatically handled.
Alternatively, you can use the BlockRebalanceOnPoll
option in
combination with AllowRebalance
to ensure rebalance cannot happen after
you poll until you explicitly allow it. This option is much easier to reason
about, but has a risk if processing your poll takes so long that a rebalance
started and finished and you were kicked from the group. If you use this option
and this API, it is recommended to take care and use PollRecords
and
ensure your RebalanceTimeout
is long enough to encompass any processing you
do between polls. My recommendation is to block rebalance on poll, ensure your
processing is quick, and to use CommitUncommittedOffsets
.
You can use the ConsumePartitions
option to assign partitions manually and
consume outside of the context of a group. If you want to use Kafka to manage
group offsets even with direct partition assignment, this repo provides a
kadm
package to easily manage offsets via an admin interface. Check the
manual_committing
example to see some
example code for how to do this.
There are two easy patterns to success for offset management in a normal consumer group.
First and the most recommended option, you can just rely on the default autocommitting behavior and the default blocking commit on leave. At most, you may want to use your own custom commit callback.
Alternatively, you can disable autocommitting with DisableAutoCommit
and instead use a custom OnPartitionsRevoked
.
In your custom revoke, you can guard a revoked variable with a mutex. Before
committing, check this revoked variable and do not commit if it has been set.
For some hints as to how to do this properly, check how
GroupTransactSession
is implemented (albeit it is more complicated due
to handling transactions).
Because an EOS consumer is difficult to implement correctly, all details have
been abstracted away to a GroupTransactSession
type. See the
transactions page for more details.
Kafka 2.4.0 introduced support for KIP-429, the incremental rebalancing protocol. This allows consumers to continue fetching records during a rebalance, effectively eliminating the stop the world aspect of rebalancing. However, while the Java client introduced this in Kafka 2.4.0, the balancer is not actually dependent on that Kafka version.
This client has support for KIP-429 and in fact defaults to cooperative consuming. Cooperative consuming is not compatible with clients using the historical consumer group strategies, and if you plan to use kgo with these historical clients, you need to set the balancers appropriately.
Cooperative rebalancing allows a client to continue fetching during rebalances, even during transactions. For transactions, a transact session will only be aborted if the member has partitions revoked.
Kafka 2.4.0 also introduced support for KIP-345, the "static" member concept for consumer group members. This is a relatively simple concept that basically just means that group members must be managed out of band from the client, whereas historically, member IDs were newly determined every time a client connected.
Static membership avoids unnecessary partition migration during rebalances and
conveys a host of other benefits; see the KIP for more details. To use static
membership, your cluster must be at least 2.4.0, and you can use the
InstanceID
option.