32 results found
-
Ease the setup of Kafka Connectors when using internal Aiven services
As a developer,
I want to have an easy setup of my Kafka connectors when they are using internal Aiven services such as Postgres or OpenSearch
so that I can save time, avoid mistakes, and enforce the strength of the platform.2 votes -
Improve support for Debezium SQL Server use cases
As an architect, I want to bring business logic out of the database and into a decoupled stream processing / event driven architecture framework. With SQL Server, I want to stream changes to Apache Kafka, using Debezium. This must support use cases where sensitivity classifications in SQL Server are used/required for things like PII.
2 votes -
Send a user alert in case stale credentials are being detected in connector configuration for Aiven for Apache Kafka Connect.
As a developer,
I want to be notified when I have stale credentials in my connectors,
so that I can update them and make sure connection from a connector to a source or sink system could be successfully established.2 votes -
Debezium Connector for Cassandra
As a developer,
I want to connect to my Cassandra databases,
so that I can use Change Data Capture (CDC) pattern to be notified about data changes and can ingest that data in to my data streaming pipeline for further use and processing.2 votes -
Exactly-Once support in Storage Write API from our GBQ sink connector
Add support for exactly-once delivery in Storage Write API for GBQ sink connector.
1 vote -
Aiven's S3 sink connector - Configure `offset.flush.interval.ms` on the connector's level
As a developer that uses Aiven's S3 sink connector,
I want to be able to set theoffset.flush.interval.ms
only for my specific connector from the connector's configuration
so that I can avoid configuring it in the cluster level (for all connectors).1 vote -
Event log should show when a connector is paused and resumed.
As a developer
I want to know when a connector is paused or resumed
so that I can have timestamps and know if anybody is doing what they are not supposed to do.1 vote -
Decompression transform for Connectors
As an application developer
I want to compress my kafka messages, but be able to decomrpess them using a transform before sinking them into a destination
so that I can save on storage costs
In addition, I'd like to use ZTSD, but more common libraries might be enough.Note, Confluent has something similar :
https://docs.confluent.io/platform/current/connect/transforms/gzipdecompress.html1 vote -
io.debezium.transforms.partitions.PartitionRouting
It would be great it the
io.debezium.transforms.partitions.PartitionRouting
SMT was available for use when configuring a Kafka Debezium source connector on the Aiven platform.1 vote -
Support for schemaless JSON messages in the Big Query Sink connector
As a developer
I want to be able to publish schemaless JSON messages to Big Query and have the BigQuery schema be updated to reflect those changes
so that I can evolve my message schema without breaking my pipeline.1 vote -
Rockset Sink connector for Kafka Connect
As a data engineer,
I want to write data to Rockset,
so that I can run my analytical workloads.1 vote -
Apache NiFi Source and Sink Kafka connector
As a developer
I want to use Apache NiFi Kafka Connector
so that I can move data in and out Apache NiFi to orchestrate data flow.1 vote
- Don't see your idea?