Awesome
A Thin Scala Wrapper Around the Kafka Streams Java API
The library wraps Java APIs in Scala thereby providing:
- much better type inference in Scala
- less boilerplate in application code
- the usual builder-style composition that developers get with the original Java API
- complete compile time type safety
The design of the library was inspired by the work started by Alexis Seigneurin in this repository.
Quick Start
kafka-streams-scala
is published and cross-built for Scala 2.11
, and 2.12
, so you can just add the following to your build:
val kafka_streams_scala_version = "0.2.1"
libraryDependencies ++= Seq("com.lightbend" %%
"kafka-streams-scala" % kafka_streams_scala_version)
Note:
kafka-streams-scala
supports onwards Kafka Streams1.0.0
.
The API docs for kafka-streams-scala
is available here for Scala 2.12 and here for Scala 2.11.
Running the Tests
The library comes with an embedded Kafka server. To run the tests, simply run sbt testOnly
and all tests will run on the local embedded server.
The embedded server is started and stopped for every test and takes quite a bit of resources. Hence it's recommended that you allocate more heap space to
sbt
when running the tests. e.g.sbt -mem 2000
.
$ sbt -mem 2000
> +clean
> +test
Type Inference and Composition
Here's a sample code fragment using the Scala wrapper library. Compare this with the Scala code from the same example in Confluent's repository.
// Compute the total per region by summing the individual click counts per region.
val clicksPerRegion: KTableS[String, Long] = userClicksStream
// Join the stream against the table.
.leftJoin(userRegionsTable, (clicks: Long, region: String) => (if (region == null) "UNKNOWN" else region, clicks))
// Change the stream from <user> -> <region, clicks> to <region> -> <clicks>
.map((_, regionWithClicks) => regionWithClicks)
// Compute the total per region by summing the individual click counts per region.
.groupByKey
.reduce(_ + _)
Implicit Serdes
One of the areas where the Java APIs' verbosity can be reduced is through a succinct way to pass serializers and de-serializers to the various functions. The library uses the power of Scala implicits towards this end. The library makes some decisions that help implement more succinct serdes in a type safe manner:
- No use of configuration based default serdes. Java APIs allow the user to define default key and value serdes as part of the configuration. This configuration, being implemented as
java.util.Properties
is type-unsafe and hence can result in runtime errors in case the user misses any of the serdes to be specified or plugs in an incorrect serde.kafka-streams-scala
makes this completely type-safe by allowing all serdes to be specified through Scala implicits. - The libraty offers implicit conversions from serdes to
Serialized
,Produced
,Consumed
orJoined
. Hence as a user you just have to pass in the implicit serde and all conversions toSerialized
,Produced
,Consumed
orJoined
will be taken care of automatically.
Default Serdes
The library offers a module that contains all the default serdes for the primitives. Importing the object will bring in scope all such primitives and helps reduce implicit hell.
object DefaultSerdes {
implicit val stringSerde: Serde[String] = Serdes.String()
implicit val longSerde: Serde[Long] = Serdes.Long().asInstanceOf[Serde[Long]]
implicit val byteArraySerde: Serde[Array[Byte]] = Serdes.ByteArray()
implicit val bytesSerde: Serde[org.apache.kafka.common.utils.Bytes] = Serdes.Bytes()
implicit val floatSerde: Serde[Float] = Serdes.Float().asInstanceOf[Serde[Float]]
implicit val doubleSerde: Serde[Double] = Serdes.Double().asInstanceOf[Serde[Double]]
implicit val integerSerde: Serde[Int] = Serdes.Integer().asInstanceOf[Serde[Int]]
}
Compile time typesafe
Not only the serdes, but DefaultSerdes
also brings into scope implicit Serialized
, Produced
, Consumed
and Joined
instances. So all APIs that accept Serialized
, Produced
, Consumed
or Joined
will get these instances automatically with an import DefaultSerdes._
.
Just one import of DefaultSerdes._
and the following code does not need a bit of Serialized
, Produced
, Consumed
or Joined
to be specified explicitly or through the default config. And the best part is that for any missing instances of these you get a compilation error. ..
import DefaultSerdes._
val clicksPerRegion: KTableS[String, Long] =
userClicksStream
// Join the stream against the table.
.leftJoin(userRegionsTable, (clicks: Long, region: String) => (if (region == null) "UNKNOWN" else region, clicks))
// Change the stream from <user> -> <region, clicks> to <region> -> <clicks>
.map((_, regionWithClicks) => regionWithClicks)
// Compute the total per region by summing the individual click counts per region.
.groupByKey
.reduce(_ + _)
// Write the (continuously updating) results to the output topic.
clicksPerRegion.toStream.to(outputTopic)