Releases: redpanda-data/connect
v4.10.0
For installation instructions check out the getting started guide.
Added
- The
nats_jetstream
input now adds a range of useful metadata information to messages. - Field
transaction_type
added to theazure_table_storage
output, which deprecates the previousinsert_type
field and supports interpolation functions. - Field
logged_batch
added to thecassandra
output. - All
sql
components now support Snowflake. - New
azure_table_storage
input. - New
sql_raw
input. - New
tracing_id
bloblang function. - New
with
bloblang method. - Field
multi_header
added to thekafka
andkafka_franz
inputs. - New
cassandra
input. - New
base64_encode
andbase64_decode
functions for the awk processor. - Param
use_number
added to theparse_json
bloblang method. - Fields
init_statement
andinit_files
added to all sql components. - New
find
andfind_all
bloblang array methods.
Fixed
- The
gcp_cloud_storage
output no longer ignores errors when closing a written file, this was masking issues when the target bucket was invalid. - Upgraded the
kafka_franz
input and output to use github.com/twmb/[email protected] since some bug fixes were made recently. - Fixed an issue where a
read_until
child input with processors affiliated would block graceful termination. - The
--labels
linting option no longer flags resource components.
The full change log can be found here.
v4.9.1
For installation instructions check out the getting started guide.
Added
- Go API: A new
BatchError
type added for distinguishing errors of a given batch.
Fixed
- Rolled back
kafka
input and output underlying sarama client library to fix a regression introduced in 4.9.0 😅 whereinvalid configuration (Consumer.Group.Rebalance.GroupStrategies and Consumer.Group.Rebalance.Strategy cannot be set at the same time)
errors would prevent consumption under certain configurations. We've decided to roll back rather than upgrade as a breaking API change was introduced that could cause issues for Go API importers (more info here: IBM/sarama#2358).
The full change log can be found here.
v4.9.0
For installation instructions check out the getting started guide.
Added
- New
parquet
input for reading a batch of Parquet files from disk. - Field
max_in_flight
added to theredis_list
input.
Fixed
- Upgraded
kafka
input and output underlying sarama client library to fix a regression introduced in 4.7.0 whereThe requested offset is outside the range of offsets maintained by the server for the given topic/partition
errors would prevent consumption of partitions. - The
cassandra
output now inserts logged batches of data rather than the less efficient (and unnecessary) unlogged form.
The full change log can be found here.
v4.8.0
For installation instructions check out the getting started guide.
Added
- All
sql
components now support Oracle DB.
Fixed
- All SQL components now accept an empty or unspecified
args_mapping
as an alias for no arguments. - Field
unsafe_dynamic_query
added to thesql_raw
output. - Fixed a regression in 4.7.0 where HTTP client components were sending duplicate request headers.
The full change log can be found here.
v4.7.0
For installation instructions check out the getting started guide.
Added
- Field
avro_raw_json
added to theschema_registry_decode
processor. - Field
priority
added to thegcp_bigquery_select
input. - The
hash
bloblang method now supportscrc32
. - New
tracing_span
bloblang function. - All
sql
components now support SQLite. - New
beanstalkd
input and output. - Field
json_marshal_mode
added to themongodb
input. - The
schema_registry_encode
andschema_registry_decode
processors now support Basic, OAuth and JWT authentication.
Fixed
- The streams mode
/ready
endpoint no longer returns status503
for streams that gracefully finished. - The performance of the bloblang
.explode
method now scales linearly with the target size. - The
influxdb
andlogger
metrics outputs should no longer mix up tag names. - Fix a potential race condition in the
read_until
connect check on terminated input. - The
parse_parquet
bloblang method andparquet_decode
processor now automatically parseBYTE_ARRAY
values as strings when the logical type is UTF8. - The
gcp_cloud_storage
output now correctly cleans up temporary files on error conditions when the collision mode is set to append.
The full change log can be found here.
v4.6.0
For installation instructions check out the getting started guide.
Added
- New
squash
bloblang method. - New top-level config field
shutdown_delay
for delaying graceful termination. - New
snowflake_id
bloblang function. - Field
wait_time_seconds
added to theaws_sqs
input. - New
json_path
bloblang method. - New
file_json_contains
predicate for unit tests. - The
parquet_encode
processor now supports theUTF8
logical type for columns.
Fixed
- The
schema_registry_encode
processor now correctly assumes Avro JSON encoded documents by default. - The
redis
processorretry_period
no longer shows linting errors for duration strings. - The
/inputs
and/outputs
endpoints for dynamic inputs and outputs now correctly render configs, both structured within the JSON response and the raw config string. - Go API: The stream builder no longer ignores
http
configuration. Instead, the value ofhttp.enabled
is set tofalse
by default.
The full change log can be found here.
v4.5.1
For installation instructions check out the getting started guide.
Fixed
- Reverted
kafka_franz
dependency back to1.3.1
due to a regression in TLS/SASL commit retention. - Fixed an unintentional linting error when using interpolation functions in the
elasticsearch
outputsaction
field.
The full change log can be found here.
v4.5.0
For installation instructions check out the getting started guide.
Added
- Field
batch_size
added to thegenerate
input. - The
amqp_0_9
output now supports setting thetimeout
of publish. - New experimental input codec
avro-ocf:marshaler=x
. - New
mapping
andmutation
processors. - New
parse_form_url_encoded
bloblang method. - The
amqp_0_9
input now supports setting theauto-delete
bit during queue declaration. - New
open_telemetry_collector
tracer. - The
kafka_franz
input and output now supports no-op SASL options with the mechanismnone
. - Field
content_type
added to thegcp_cloud_storage
cache.
Fixed
- The
mongodb
processor and output defaultwrite_concern.w_timeout
empty value no longer causes configuration issues. - Field
message_name
added to the logger config. - The
amqp_1
input and output should no longer spam logs with timeout errors during graceful termination. - Fixed a potential crash when the
contains
bloblang method was used to compare complex types. - Fixed an issue where the
kafka_franz
input or output wouldn't use TLS connections without custom certificate configuration. - Fixed structural cycle in the CUE representation of the
retry
output. - Tracing headers from HTTP requests to the
http_server
input are now correctly extracted.
Changed
- The
broker
input no longer applies processors before batching as this was unintentional behaviour and counter to documentation. Users that rely on this behaviour are advised to place their pre-batching processors at the level of the child inputs of the broker. - The
broker
output no longer applies processors after batching as this was unintentional behaviour and counter to documentation. Users that rely on this behaviour are advised to place their post-batching processors at the level of the child outputs of the broker.
The full change log can be found here.
v4.5.0-rc1
For installation instructions check out the getting started guide.
Added
- Field
batch_size
added to thegenerate
input. - The
amqp_0_9
output now supports setting thetimeout
of publish. - New experimental input codec
avro-ocf:marshaler=x
. - New
mapping
andmutation
processors. - New
parse_form_url_encoded
bloblang method. - The
amqp_0_9
input now supports setting theauto-delete
bit during queue declaration. - New
open_telemetry_collector
tracer.
Fixed
- The
mongodb
processor and output defaultwrite_concern.w_timeout
empty value no longer causes configuration issues. - Field
message_name
added to the logger config. - The
amqp_1
input and output should no longer spam logs with timeout errors during graceful termination. - Fixed a potential crash when the
contains
bloblang method was used to compare complex types. - Fixed an issue where the
kafka_franz
input or output wouldn't use TLS connections without custom certificate configuration. - Fixed structural cycle in the CUE representation of the
retry
output.
Changed
- The
broker
input no longer applies processors before batching as this was unintentional behaviour and counter to documentation. Users that rely on this behaviour are advised to place their pre-batching processors at the level of the child inputs of the broker. - The
broker
output no longer applies processors after batching as this was unintentional behaviour and counter to documentation. Users that rely on this behaviour are advised to place their post-batching processors at the level of the child outputs of the broker.
The full change log can be found here.
v4.4.1
For installation instructions check out the getting started guide.
Fixed
- Fixed an issue where an
http_server
input or output would fail to register prometheus metrics when combined with other inputs/outputs. - Fixed an issue where the
jaeger
tracer was incapable of sending traces to agents outside of the default port.
The full change log can be found here.