You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: source/kafka-source.txt
+112-3Lines changed: 112 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -126,19 +126,21 @@ an example source connector configuration file, see
126
126
* - database
127
127
- string
128
128
- | Name of the database to watch for changes. If not set, all databases are watched.
129
+
|
129
130
| **Default**: ""
130
131
| **Accepted Values**: A single database name
131
132
132
133
* - collection
133
134
- string
134
-
- | Name of the collection in the database to watch for changes.
135
-
| The collection in the database to watch. If not set then all collections will be watched.
135
+
- | Name of the collection in the database to watch for changes. If not set then all collections will be watched.
136
+
|
136
137
| **Default**: ""
137
138
| **Accepted Values**: A single collection name
138
139
139
140
* - publish.full.document.only
140
141
- boolean
141
142
- | Only publish the changed document instead of the full change stream document. Sets the ``change.stream.full.document=updateLookup`` automatically so updated documents will be included.
143
+
|
142
144
| **Default**: false
143
145
| **Accepted Values**: ``true`` or ``false``
144
146
@@ -162,58 +164,131 @@ an example source connector configuration file, see
162
164
* - collation
163
165
- string
164
166
- | A JSON :manual:`collation document </reference/collation/#collation-document>` that contains options to use for the change stream. Append ``.asDocument().toJson()`` to the collation document to create the JSON representation.
167
+
|
165
168
| **Default**: ""
166
169
| **Accepted Values**: A valid JSON document representing a collection
167
170
168
171
* - output.format.key
169
172
- string
170
173
- | Determines which data format the source connector outputs for the key document.
- | Whether the connector should infer the schema for the value. Since each document is processed in isolation, multiple schemas may result. Only valid when ``schema`` is specified in the ``output.format.value`` setting.
256
+
|
193
257
| **Default**: ``false``
194
258
| **Accepted Values**: ``true`` or ``false``
195
259
260
+
* - offset.partition.name
261
+
- string
262
+
- | Custom partition name to use in which to store the offset values. The offset value stores information on where to resume processing if there is an issue that requires you to restart the connector. By choosing a new partition name, you can start processing without using a resume token. This can make it easier to restart the connector without reconfiguring the Kafka Connect service or manually deleting the old offset. The offset partition is automatically created if it does not exist.
263
+
|
264
+
| **Default**: ""
265
+
| **Accepted Values**: A string
266
+
196
267
* - batch.size
197
268
- int
198
269
- | The cursor batch size.
270
+
|
199
271
| **Default**: 0
200
272
| **Accepted Values**: An integer
201
273
202
274
* - change.stream.full.document
203
275
- string
204
276
- | Determines what to return for update operations when using a Change Stream. When set to 'updateLookup', the change stream for partial updates will include both a delta describing the changes to the document as well as a copy of the entire document that was changed from *some point in time* after the change occurred.
277
+
|
205
278
| **Default**: ""
206
279
| **Accepted Values**: "" or ``default`` or ``updateLookup``
207
280
208
281
* - poll.await.time.ms
209
282
- long
210
283
- | The amount of time to wait before checking for new results on the change stream
284
+
|
211
285
| **Default**: 5000
212
286
| **Accepted Values**: An integer
213
287
214
288
* - poll.max.batch.size
215
289
- int
216
290
- | Maximum number of change stream documents to include in a single batch when polling for new data. This setting can be used to limit the amount of data buffered internally in the connector.
291
+
|
217
292
| **Default**: 1000
218
293
| **Accepted Values**: An integer
219
294
@@ -229,12 +304,32 @@ an example source connector configuration file, see
229
304
* - copy.existing
230
305
- boolean
231
306
- | Copy existing data from source collections and convert them to Change Stream events on their respective topics. Any changes to the data that occur during the copy process are applied once the copy is completed.
307
+
|
232
308
| **Default**: false
233
309
| **Accepted Values**: ``true`` or ``false``
234
310
311
+
* - copy.existing.namespace.regex
312
+
- string
313
+
- | Regular expression that matches the namespaces from which to copy
314
+
data. A namespace describes the database name and collection
315
+
separated by a period, e.g. ``databaseName.collectionName``.
316
+
317
+
.. example::
318
+
319
+
In the following example, the setting matches all collections
320
+
that start with "page" in the "stats" database.
321
+
322
+
.. code-block:: none
323
+
324
+
copy.existing.namespace.regex=stats\.page.*
325
+
326
+
| **Default**: ""
327
+
| **Accepted Values**: A valid regular expression
328
+
235
329
* - copy.existing.max.threads
236
330
- int
237
331
- | The number of threads to use when performing the data copy. Defaults to the number of processors.
332
+
|
238
333
| **Default**: defaults to the number of processors
239
334
| **Accepted Values**: An integer
240
335
@@ -307,6 +402,20 @@ an example source connector configuration file, see
307
402
| **Default:** ""
308
403
| **Accepted Values**: A valid partition name
309
404
405
+
* - heartbeat.interval.ms
406
+
- int
407
+
- | The length of time in milliseconds between sending heartbeat messages to record a post batch resume token when no source records have been published. This can improve the resumability of the connector for low volume namespaces. Use ``0`` to disable.
408
+
|
409
+
| **Default**: ``0``
410
+
| **Accepted Values**: An integer
411
+
412
+
* - heartbeat.topic.name
413
+
- string
414
+
- | The name of the topic to write heartbeat messages to.
415
+
|
416
+
| **Default**: ``__mongodb_heartbeats``
417
+
| **Accepted Values**: A valid Kafka topic name
418
+
310
419
.. note::
311
420
312
421
The default maximum size for Kafka messages is 1MB. Update the
0 commit comments