Confluent スキーマレジストリ

Ensure Application Development Compatibility

Confluent Schema Registryは、開発者がイベントの標準スキーマを定義して組織内で共有し、前方/後方互換性を担保しながらスキーマを安全に進化させるための RESTful インターフェイスを提供します。

Download Confluent
Schema Registry hero image

Why Confluent Schema Registry?

Schema Registry stores a versioned history of all schemas and allows the evolution of schemas according to the configured compatibility settings. It also provides a plugin to clients that handles schema storage and retrieval for messages that are sent in Avro format.

Deploy Reliably

Let your developers focus on deploying applications freely while protecting your Apache Kafka deployment from breaking changes.

Evolve Quickly

Need to add a new column to a downstream database? You don’t need an involved change process and at least 5 meetings to coordinate 20 teams.

Validate your changes as an integrated part of the development process, reducing coordination overhead and ensuring information integrity in the environment.

How Can Confluent Schema Registry Help You?

Schema Registry Learn to love image

Manage your Schema

Schemas are named and you can have multiple versions of the same schema. Schema Registry validates compatibility and warns about possible issues. This allows developers to add and remove fields independently to move faster. New schemas and versions are automatically registered and automatically validated, making the pushing to production process seamless.

Integrate with Standard Development Tools

To help you find out about compatibility issues as early as possible, Schema Registry includes Maven plugins, which enables integration of schema management and validation right into the development process.

Access your Schema

Schema Registry provides a REST API that allows any application to integrate and save or retrieve schemas. Additionally, formatters provide command line functionality for automatically converting JSON messages to make your data human friendly.

Provide Automated Serialization

Confluent Schema Registry is built right into Kafka Client serialization libraries—in every language. Writing serialized Apache Avro records to Kafka is as simple as configuring a producer with the Schema Registry serializers and sending Avro objects to Kafka. And if you use incompatible schema accidentally? That’s just an exception for the producer to handle. Incompatible data will never make it into Apache Kafka.

How Confluent Schema Registry Works

How it works image

The serializer places a call to the schema registry, to see if it has a format for the data the application wants to publish. If it does, schema registry passes that format to the application’s serializer, which uses it to filter out incorrectly formatted messages.


After checking the schema is authorized, it’s automatically serialized and there’s no effort you need to put into it. The message will, as expected, be delivered to the Kafka topic.


Your consumers will handle deserialization, making sure your data pipeline can quickly evolve and continue to have clean data. You simply need to have all applications call the schema registry when publishing.

Get started or get to know us better

We use cookies to understand how you use our site and to improve your experience. Click here to learn more or change your cookie settings. By continuing to browse, you agree to our use of cookies.