Copyright 2018 ABSA Group Limited
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
Pain free Spark/Avro integration.
Seamlessly convert your Avro records from anywhere (e.g. Kafka, Parquet, HDFS, etc) into Spark Rows.
Convert your Dataframes into Avro records without even specifying a schema.
Seamlessly integrate with Confluent platform, including Schema Registry.
Among the motivations for this project, it is possible to highlight:
-
Avro is Confluent's recommended payload to Kafka (https://www.confluent.io/blog/avro-kafka-data/).
-
Current solutions do not support reading Avro record as Spark structured streams (e.g. https://github.com/databricks/spark-avro).
-
Lack tools for writing Spark Dataframes directly into Kafka as Avro records.
-
Right now, all the efforts to integrated Avro-Spark streams or Dataframes depend on the user.
-
Confluent's Avro payload requires specialized deserializers, since they send schema metadata with it, which makes the integration with Spark cumbersome.
-
In some use cases you may want to keep your Dataframe schema, adding your data as a nested structure, whereas in other cases you may want to use the schema for your data as the schema for the whole Dataframe.
On spark 2.4.x Abris should work without any further requirements.
On Spark 2.3.x you must declare dependency on org.apache.avro:avro:1.8.0
or higher. (Spark 2.3.x uses Avro 1.7.x so you must overwrite this because Abris needs Avro 1.8.0+.)
ABRiS API is in it's most basic form almost identical to spark built-in support for Avro, but it provides additional functionality. Mainly it's support of schema registry and also seamless integration with confluent Avro data format.
The API consists of four Spark SQL expressions:
to_avro
andfrom_avro
used for normal Avro payloadto_confluent_avro
andfrom_confluent_avro
used for Confluent Avro data format
Full runnable examples can be found in za.co.absa.abris.examples
package. You can also take a look at unit tests in package za.co.absa.abris.avro.sql
.
Old ABRiS API is deprecated, but is still included in the library. Documentation for old API is in ABRiS 2.2.3.
The format of Avro binary data is defined in Avro specification. Confluent format extends it and prepends the schema id before the rest of binary data. The Confluent expressions in this library expect this format and add the id after the Avro data are generated or remove it before they are parsed.
You can find more about confluent and Schema Registry in Confluent documentation.
import za.co.absa.abris.avro.functions.from_avro
def readAvro(dataFrame: DataFrame, schemaString: String): DataFrame = {
dataFrame.select(from_avro(col("value"), schemaString) as 'data).select("data.*")
}
In this example the Avro binary data are in dataFrame
inside column named value. The Avro schema is provided as a string schemaString
.
After the Avro data are converted to Spark SQL representation they are stored in column named data. This column is immediately flattened in the next select so the result will be a DataFrame
containing only the deserialized avro data.
If you want to use Schema Registry you need to provide a configuration:
val schemaRegistryConfig = Map(
SchemaManager.PARAM_SCHEMA_REGISTRY_URL -> "url_to_schema_registry",
SchemaManager.PARAM_SCHEMA_REGISTRY_TOPIC -> "topic_name",
SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY -> SchemaManager.SchemaStorageNamingStrategies.{TOPIC_NAME, RECORD_NAME, TOPIC_RECORD_NAME}, // choose a subject name strategy
SchemaManager.PARAM_VALUE_SCHEMA_ID -> "current_schema_id" // set to "latest" if you want the latest schema version to used
)
Depending on selected naming strategy you may also need to provide SchemaManager.PARAM_SCHEMA_NAME_FOR_RECORD_STRATEGY
and SchemaManager.PARAM_SCHEMA_NAMESPACE_FOR_RECORD_STRATEGY
import za.co.absa.abris.avro.functions.from_avro
def readAvro(dataFrame: DataFrame, schemaRegistryConfig: Map[String, String]): DataFrame = {
dataFrame.select(from_avro(col("value"), schemaRegistryConfig) as 'data).select("data.*")
}
This part is almost same as in previous example only this time you send in the registry configuration instead of schema. Of course the schema already have to be in Schema Registry for this to work.
import za.co.absa.abris.avro.functions.from_confluent_avro
def readAvro(dataFrame: DataFrame, schemaString: String): DataFrame = {
dataFrame.select(from_confluent_avro(col("value"), schemaString) as 'data).select("data.*")
}
The main difference between from_confluent_avro
and from_avro
is in whether it expects the schema_id in the Avro payload. The usage is identical to previous examples.
Schema Registry configuration is the same as in previous Schema Registry example.
import za.co.absa.abris.avro.functions.from_confluent_avro
def readAvro(dataFrame: DataFrame, schemaRegistryConfig: Map[String, String]): DataFrame = {
dataFrame.select(from_confluent_avro(col("value"), schemaRegistryConfig) as 'data).select("data.*")
}
The only difference is the expression name.
In a case that we are sending the Avro data using Kafka we may want to serialize both the key and the value of Kafka message. The serialization of Avro data is not really different when we are doing it for key or for value, but Schema Registry handle each of them slightly differently. Specifically the subject may depend on this.
The way the library knows whether you are working with key or value is the schema naming strategy.
Use SchemaManager.PARAM_KEY_SCHEMA_NAMING_STRATEGY
for a key and SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY
for a value. If the configuration will contain both of them it will throw!
This is one way how to create the configurations for key and value serialization:
val commonRegistryConfig = Map(
SchemaManager.PARAM_SCHEMA_REGISTRY_TOPIC -> "example_topic",
SchemaManager.PARAM_SCHEMA_REGISTRY_URL -> "http://localhost:8081"
)
val keyRegistryConfig = commonRegistryConfig ++ Map(
SchemaManager.PARAM_KEY_SCHEMA_NAMING_STRATEGY -> "topic.record.name",
SchemaManager.PARAM_KEY_SCHEMA_ID -> "latest",
SchemaManager.PARAM_SCHEMA_NAME_FOR_RECORD_STRATEGY -> "foo",
SchemaManager.PARAM_SCHEMA_NAMESPACE_FOR_RECORD_STRATEGY -> "com.bar"
)
val valueRegistryConfig = commonRegistryConfig ++ Map(
SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY -> "topic.name",
SchemaManager.PARAM_VALUE_SCHEMA_ID -> "latest"
)
Let's assume that the Avro binary data for key are in key
column and the value data are in value
column of the same DataFrame.
import za.co.absa.abris.avro.functions.from_confluent_avro
val result: DataFrame = dataFrame.select(
from_confluent_avro(col("key"), keyRegistryConfig) as 'key,
from_confluent_avro(col("value"), valueRegistryConfig) as 'value)
We just need to use the right configuration for the right column and that's it.
import za.co.absa.abris.avro.functions.to_avro
def writeAvro(dataFrame: DataFrame): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_avro(allColumns) as 'value)
}
This is the simplest possible usage of to_avro
expression. We just provide the column that we want to serialize and the library will generate the schema automatically from spark data types.
If you want to serialize more than one column you have to put them in a spark struct first, as you can see in the example.
import za.co.absa.abris.avro.functions.to_avro
def writeAvro(dataFrame: DataFrame, schemaString: String): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_avro(allColumns, schemaString) as 'value)
}
If you provide the Avro schema as a second argument the library will use it to convert Spark data into Avro. If the data types in Spark DataFrame and in schema aren't compatible it may cause problems.
First we need to provide Schema Registry configuration:
val schemaRegistryConfig = Map(
SchemaManager.PARAM_SCHEMA_REGISTRY_URL -> "url_to_schema_registry",
SchemaManager.PARAM_SCHEMA_REGISTRY_TOPIC -> "topic_name",
SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY -> SchemaManager.SchemaStorageNamingStrategies.TOPIC_RECORD_NAME,
SchemaManager.PARAM_SCHEMA_NAME_FOR_RECORD_STRATEGY -> "schema_name",
SchemaManager.PARAM_SCHEMA_NAMESPACE_FOR_RECORD_STRATEGY -> "schema_namespace"
)
In this example the TOPIC_RECORD_NAME
naming strategy is used therefore we need to provide topic, name and namespace.
import za.co.absa.abris.avro.functions.to_avro
def writeAvro(dataFrame: DataFrame, schemaRegistryConfig: Map[String, String]): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_avro(allColumns, schemaRegistryConfig) as 'value)
}
Since we didn't provide a schema it will be generated automatically and than stored in the Schema Registry.
The only difference from previous example is that we have one additional parameter for the schema.
import za.co.absa.abris.avro.functions.to_avro
def writeAvro(dataFrame: DataFrame, schemaString: String, schemaRegistryConfig: Map[String, String]): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_avro(allColumns, schemaString, schemaRegistryConfig) as 'value)
}
Schema Registry configuration is the same as in previous examples.
import za.co.absa.abris.avro.functions.to_confluent_avro
def writeAvro(dataFrame: DataFrame, schemaRegistryConfig: Map[String, String]): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_confluent_avro(allColumns, schemaRegistryConfig) as 'value)
}
The main difference between from_confluent_avro
and from_avro
is in whether it prepends the schema_id in the Avro payload. The usage is identical to previous examples.
import za.co.absa.abris.avro.functions.to_confluent_avro
def writeAvro(dataFrame: DataFrame, schemaString: String, schemaRegistryConfig: Map[String, String]): DataFrame = {
val allColumns = struct(dataFrame.columns.head, dataFrame.columns.tail: _*)
dataFrame.select(to_confluent_avro(allColumns, schemaString, registryConfig) as 'value)
}
As in reading example, for writing you also have to specify the naming strategy parameter to let ABRiS know if you are using value or key.
val commonRegistryConfig = Map(
SchemaManager.PARAM_SCHEMA_REGISTRY_TOPIC -> "example_topic",
SchemaManager.PARAM_SCHEMA_REGISTRY_URL -> "http://localhost:8081",
SchemaManager.PARAM_SCHEMA_NAME_FOR_RECORD_STRATEGY -> "foo",
SchemaManager.PARAM_SCHEMA_NAMESPACE_FOR_RECORD_STRATEGY -> "com.bar"
)
val keyRegistryConfig = commonRegistryConfig +
(SchemaManager.PARAM_KEY_SCHEMA_NAMING_STRATEGY -> "topic.record.name")
val valueRegistryConfig = commonRegistryConfig +
(SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY -> "topic.name")
Let's assume that we have the data that we want to serialize in a DataFrame in key and value columns.
val result: DataFrame = dataFrame.select(
to_confluent_avro(col("key"), keyRegistryConfig) as 'key,
to_confluent_avro(col("value"), valueRegistryConfig) as 'value)
After serialization the data are again stored in columns key and value, but now they are in Avro binary format.
This library provides utility methods for registering schemas with topics into Schema Registry. Below is an example of how it can be done.
val schemaRegistryConfs = Map(
SchemaManager.PARAM_SCHEMA_REGISTRY_URL -> "url_to_schema_registry",
SchemaManager.PARAM_VALUE_SCHEMA_NAMING_STRATEGY -> SchemaManager.SchemaStorageNamingStrategies.{TOPIC_NAME, RECORD_NAME, TOPIC_RECORD_NAME}, // if you are retrieving value schema
SchemaManager.PARAM_KEY_SCHEMA_NAMING_STRATEGY -> SchemaManager.SchemaStorageNamingStrategies.{TOPIC_NAME, RECORD_NAME, TOPIC_RECORD_NAME}, // if your are retrieving key schema
SchemaManager.PARAM_SCHEMA_NAME_FOR_RECORD_STRATEGY -> "schema_name", // if you're using RecordName or TopicRecordName strategies
SchemaManager.PARAM_SCHEMA_NAMESPACE_FOR_RECORD_STRATEGY -> "schema_namespace" // if you're using RecordName or TopicRecordName strategies
)
SchemaManager.configureSchemaRegistry(schemaRegistryConfs)
val topic = "example_topic"
val subject = SchemaManager.getSubjectName(topic, false) // create a subject for the value
val schema = AvroSchemaUtils.load("path_to_the_schema_in_a_file_system")
val schemaId = SchemaManager.register(schema, subject)
This library also provides convenient methods to convert between Avro and Spark schemas.
If you have an Avro schema which you want to convert into a Spark SQL one - to generate your Dataframes, for instance - you can do as follows:
val avroSchema: Schema = AvroSchemaUtils.load("path_to_avro_schema")
val sqlSchema: StructType = SparkAvroConversions.toSqlType(avroSchema)
You can also do the inverse operation by running:
val sqlSchema = new StructType(new StructField ....
val avroSchema = SparkAvroConversions.toAvroSchema(sqlSchema, avro_schema_name, avro_schema_namespace)
Your data may not come from Spark and, if you are using Avro, there is a chance you are also using Java. This library provides utilities to easily convert Java beans into Avro records and send them to Kafka.
To use you need to:
- Store the Avro schema for the class structure you want to send (the Spark schema will be inferred from there later, in your reading job):
Class<?> dataType = YourBean.class;
String schemaDestination = "/src/test/resources/DestinationSchema.avsc";
AvroSchemaGenerator.storeSchemaForClass(dataType, Paths.get(schemaDestination));
- Write the configuration to access your cluster:
Properties config = new Properties();
config.put("bootstrap.servers", "...
...
- Create an instance of the writer:
KafkaAvroWriter<YourBean> writer = new KafkaAvroWriter<YourBean>(config);
- Send your data to Kafka:
List<YourBean> data = ...
long dispatchWait = 1l; // how much time the writer should expect until all data are dispatched
writer.write(data, "destination_topic", dispatchWait);
A complete application can be found at za.co.absa.abris.utils.examples.SimpleAvroDataGenerator
under Java source.
The naming strategies RecordName and TopicRecordName allow a topic to receive different payloads, i.e. payloads containing different schemas that do not have to be compatible, as explained here.
However, currently, there is no way for Spark to change Dataframes schemas on the fly, thus, if incompatible schemas are used on the same topic, the job will fail. Also, it would be cumbersome to write jobs that shift between schemas.
A possible solution would be for ABRiS to create an uber schema from all schemas expected to be part of a topic, which will be investigated in future releases.
Fixed is an alternative way of encoding binary data in Avro. Unlike bytes type the fixed type doesn't store the length of the data in the payload, but in Avro schema itself.
The corresponding data type in Spark is BinaryType, but the inferred schema will always use bytes type for this kind of data. If you want to use the fixed type you must provide the Avro schema.