-
Notifications
You must be signed in to change notification settings - Fork 52
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FLINK-31989][docs] Update english docs for KinesisStreamsSource and DynamoDbStreamsSource #179
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Comments on DDB connector docs
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Many thanks for the updated docs, I've added comments inline.
<td>yes</td> | ||
<td style="word-wrap: break-word;">16</td> | ||
<td>Integer</td> | ||
<td>Request threshold for uncompleted requests by <code>KinesisAsyncClient</code>before blocking new write requests and applying backpressure.</td> |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is actually something that should be explained in DataStream too.
Also, explaining that a "request" is a batch
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In general, the batching mechanism would deserve a dedicated chapter, possibly in DataStream docs, and linked from here
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's focus the changes for this PR to the source docs.
Kinesis data streams consist of one or more shards, and the `sink.partitioner` option allows you to control how records written into a multi-shard Kinesis-backed table will be partitioned between its shards. | ||
Valid values are: | ||
|
||
* `fixed`: Kinesis `PartitionKey` values derived from the Flink subtask index, so each Flink partition ends up in at most one Kinesis partition (assuming that no re-sharding takes place at runtime). |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It would be clearer putting this are mutually exclusive options.
Partitioning is defined by either using
PARTITION BY
in the table definition or by specifying specifysink.partitioner
. Using both will result in a configuration error.
Valid values for
sink.partitioner
:
fixed
...random
...- Custom
FixedKinesisPartitioner
subclass...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's focus the changes for this PR to the source docs.
|
||
{{< hint info >}} | ||
Records written into tables defining a `PARTITION BY` clause will always be partitioned based on a concatenated projection of the `PARTITION BY` fields. | ||
In this case, the `sink.partitioner` field cannot be used to modify this behavior (attempting to do this results in a configuration error). |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If alternative options are explained above, this line becomes redundant
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's focus the changes for this PR to the source docs.
|
||
# Data Type Mapping | ||
|
||
Kinesis stores records as Base64-encoded binary data objects, so it doesn't have a notion of internal record structure. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Text formats, such as
json
orcsv
are written to Kinesis without modifications. Binary formats such asavro
are Base64-encoded and then written to Kinesis as text.
(is this right?)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not sure actually!
Let's focus the changes for this PR to the source docs.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM for the source docs
<td>no</td> | ||
<td style="word-wrap: break-word;">JOB_MANAGED</td> | ||
<td>String</td> | ||
<td>Only applicable to EFO <code>ReaderType</code>. Determine if the EFO consumer is managed by the Flink job <code>JOB_MANAGED|SELF_MANAGED</code>.</td> |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sorry, I copied&pasted the deleted title from the other docs.
It was "EFO Stream Consumer Lifecycle Management".
But we can skip linking across docs for now
Purpose of the change
Updates English docs for
KinesisStreamsSource
andDynamoDbStreamsSource
Verifying this change
This change is a docs change without any test coverage.
Significant changes
(Please check any boxes [x] if the answer is "yes". You can first publish the PR and check them afterwards, for convenience.)
@Public(Evolving)
)