MongoDB New Document State Extraction
Configuration
The configuration is a part of sink task connector and is expressed in a set of properties:
transforms=unwrap,...
transforms.unwrap.type=io.debezium.connector.mongodb.transforms.ExtractNewDocumentState
transforms.unwrap.drop.tombstones=false
transforms.unwrap.delete.handling.mode=drop
transforms.unwrap.operation.header=true
Array encoding
The SMT converts MongoDB arrays into arrays as defined by Apache Connect (or Apache Avro) schema. The problem is that such arrays must contains elements of the same type. MongoDB allows the user to store elements of heterogeneous types into the same array. To bypass this impedance mismatch it is possible to encode the array in two different ways using array.encoding
configuration option.
transforms=unwrap,...
transforms.unwrap.type=io.debezium.connector.mongodb.transforms.ExtractNewDocumentState
transforms.unwrap.array.encoding=<array|document>
Value array
(the default) will encode arrays as the array datatype. It is user’s responsibility to ensure that all elements for a given array instance are of the same time. This option is a restricting one but offers easy processing of arrays by downstream clients.
Value document
will convert the array into a struct of structs in the similar way as done by BSON serialization. The main struct contains fields named _0
, _1
, _2
etc. where the name represents the index of the element in the array. Every element is then passed as the value for the give field.
Let’s suppose an example source MongoDB document with array with heterogeneous types
{
"_id": 1,
"a1": [
{
"a": 1,
"b": "none"
},
{
"a": "c",
"d": "something"
}
]
}
This document will be encoded as
{
"_id": 1,
"a1": {
"_0": {
"a": 1,
"b": "none"
},
"_1": {
"a": "c",
"d": "something"
}
}
}
This option allows you to process arbitrary arrays but the consumer need to know how to properly handle them.
Note: The underscore in index names is present because Avro encoding requires field names not to start with digit.
Nested structure flattening
When a MongoDB document contains a nested document (structure) it is faithfully encoded as a nested structure field. If the sink connector does support only flat structure it is possible to flatten the internal structure into a flat one with a consistent field naming. To enable this feature the option flatten.struct
must be set to true
.
transforms=unwrap,...
transforms.unwrap.type=io.debezium.connector.mongodb.transforms.ExtractNewDocumentState
transforms.unwrap.flatten.struct=<true|false>
transforms.unwrap.flatten.struct.delimiter=<string>
The resulting flat document will consist of fields whose names are created by joining the name of the parent field and the name of the fields in the nested document. Those elements are separated with string defined by an option struct.delimiter
by default set to the underscore.
Let’s suppose an example source MongoDB document with a field with a nested document
{
"_id": 1,
"a": {
"b": 1,
"c": "none"
},
"d": 100
}
Such document will be encoded as
{
"_id": 1,
"a_c": 1,
"a_d": "none",
"d": 100
}
This option allows you to convert a hierarchical document into a flat structure suitable for a table-like storage.
MongoDB $unset
handling
MongoDB allows $unset
operations that remove a certain field from a document. Because the collections are schemaless, it becomes hard to inform consumers/sinkers about the field that is now missing. The approach that Debezium uses is to set the field being removed to a null value.
Given the operation
{
"after":null,
"patch":"{\"$unset\" : {\"a\" : true}}"
}
The final encoding will look like
{
"id": 1,
"a": null
}
Note that other MongoDB operations might cause an $unset
internally, $rename
is one example.
Determine original operation
When a message is flattened the final result does not show whether it was an insert, update or first read. (Deletions can be detected via tombstones or rewrites, see Configuration options.)
To solve this problem Debezium offers an option to propagate the original operation via a header added to the message. To enable this feature the option operation.header
must be set to true
.
transforms=unwrap,...
transforms.unwrap.type=io.debezium.connector.mongodb.transforms.ExtractNewDocumentState
transforms.unwrap.operation.header=true
The possible values are the ones from the op
field of MongoDB connector change events.
Adding source metadata fields
The SMT can optionally add metadata fields from the original change event’s source
structure to the final flattened record (prefixed with “__“). This functionality can be used to add things like the collection from the change event, or connector-specific fields like the replica set name. For more information on what’s available in the source structure see the documentation for the MongoDB connector.
For example, the configuration
transforms=unwrap,...
transforms.unwrap.type=io.debezium.connector.mongodb.transforms.ExtractNewDocumentState
transforms.unwrap.add.source.fields=rs,collection
will add
{ "__rs" : "rs0", "__collection" : "my-collection", ... }
to the final flattened record.
For DELETE
events, this option is only supported when the delete.handling.mode
option is set to “rewrite”.
Configuration options
Property | Default | Description |
---|---|---|
| The SMT converts MongoDB arrays into arrays as defined by Apache Connect (or Apache Avro) schema. | |
| The SMT flattens structs by concatenating the fields into plain properties, using a configurable delimiter. | |
Delimiter to concat between field names from the input record when generating field names for the output record. Only applies when | ||
| The SMT removes the tombstone generated by Debezium from the stream. | |
| The SMT can | |
(double-underscore) | Set this optional string to prefix a header. | |
Specify a list of metadata fields to add to header of the flattened message. In case of duplicate field names (e.g. “tsms” exists twice), the struct should be specified to get the correct field (e.g. “source.tsms”). The fields will be prefixed with | ||
(double-underscore) | Set this optional string to prefix a field. | |
Specify a list of metadata fields to add to the flattened message. In case of duplicate field names (e.g. “tsms” exists twice), the struct should be specified to get the correct field (e.g. “source.tsms”). The fields will be prefixed with | ||
| Whether field names will be sanitized to adhere to Avro naming requirements. See Avro naming for more details. |
Known limitations
Feeding data changes from a schemaless store such as MongoDB to strictly schema-based datastores such as a relational database can by definition work within certain limits only. Specifically, all fields of documents within one collection with the same name must be of the same type. Otherwise, no consistent column definition can be derived in the target database.
Arrays will be restored in the emitted Kafka Connect record correctly, but they are not supported by sink connector just expecting a “flat” message structure.