34 results found
-
Certificate file support for Kafka connectors
As a developer / security engineer,
I want to be able to authenticate my Apache Kafka connectors via mTLS,
so that I connect to my external services in a secure way.24 votesRoadmapped · AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded -
Azure Blob Storage Kafka connector
As a developer,
I want to read and write data to Azure Blob Storage,
so that I can use that data for stream processing and analytics, and backup / restore data in Apache Kafka.17 votesRoadmapped · AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded -
Support for AWS S3 Source Kafka Connector
As a developer / data engineer,
I want to be able to read data stored in AWS S3 bucket,
so that I can transfer, process and transform that data to other applications.
In addition, I can use data stored in S3 bucket as backup and I can rehydrate my Apache Kafka with it.16 votesRoadmapped · AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded -
Support Databricks driver for Apache Kafka Connect JDBC sink connector
As a developer,
I want to have support for Databricks driver for Apache Kafka Connect JDBC sink connector,
so that I can write data from Aiven for Apache Kafka to Databricks Spark for further processing, analysis and consumption.11 votes -
AWS SQS Source and Sink connector
As a developer,
I want to read and write data in AWS SQS
so that I can enable more architectural solutions and use SQS with other AWS services.11 votes -
Support for Apache Kafka Connect to be run standalone against a 3rd party Apache Kafka service
As a developer,
I want to run standalone Apache Kafka Connect cluster against 3rd party Apache Kafka service,
so that I can benefit from managed Apache Kafka Connect service and read/write data from/to Apache Kafka service running outside Aiven.10 votes -
ClickHouse sink for Kafka Connect
As a developer,
I want to sync my data in to Clickhouse,
so that I can store large volumes of data and run analytics on top of it.10 votes -
Add Oracle JDBC support for Aiven for Kafka Connect
As a developer,
I want to be able to connect to my Oracle database from Aiven for Apache Kafka Connect,
so that I can read and write data from/to Oracle DB and/or Apache Kafka to enable my use-cases.9 votes -
Support for "Apache Iceberg" format while sinking CDC
As a developer / DevOps
I want to be able to Sink CDC data into Apache Iceberg format
so that I can analyze data using time travel feature of AWS Athena
In addition, we may find a way for the current "Aiven - Amazon AWS S3 Sink" connector to be able to produce "Apache Iceberg" in addition of "Parquet" format or we may provide a dedicated connector like the one from this repository : https://github.com/tabular-io/iceberg-kafka-connectYours faithfully,
LCDP8 votes -
ClickHouse Driver for Apache Kafka Connect JDBC Sink connector
As a developer,
I want to have ClickHouse driver support in JDBC sink connector,
so that I can write data from Apache Kafka to ClickHouse for further processing and analysis.7 votes -
Ability to choose the Apache Kafka Connect connector version
As a developer,
I want to choose which Apache Kafka Connect connector version to use,
so that I can control connector version and make sure it is compatible with my applications.6 votes -
Salesforce Kafka Connector
As a developer,
I want to connect my Apache Kafka to Salesforce
so that I can read and write data from/to SFDC.6 votes -
Azure Cosmos DB Kafka Connector
As a developer,
I want to connect to Azure Cosmos DB from Apache Kafka,
so that I can access my data stored in Cosmos DB and use it in my stream processing pipeline.6 votes -
Aiven's S3 sink connector - Support configure offset flush max size
As Developer
I want to use Aiven's S3 sink connector and have the ability to control the size of output files (keep same size even if the traffic changes). In general, the connector has no lags and we want to flush the offset and write to file only when we have enough data. We useoffset.flush.interval.ms
for it but when the traffic increases, the amount of data arrives in the configured interval is increased and can cause an OOM issue. In addition, when we pause the connector for couple of minutes and we gather a lag, it can also lead…5 votes -
Support for Protobuf serialization of Events (and Keys) in Aiven Kafka Connnect
As a data engineer
I want Aiven Kafka Connect to offer the option to use the protocol buffer data format when serializíng the events that is sends to a Kafka broker.
In my specific case, I need it to be possible in a Debezium connector for PostgreSQL.
Additionally, it would be good to have as a user the option to define oneself the protobuf schema to use for serializing.5 votes -
Azure Data Lake Gen2 (ADLS) Kafka connector
Kafka Connector to allow for streaming data to Azure Data Lake Gen2 as a sink.
5 votes -
Apache Camel connector for Kafka Connect
As a software / data science engineer,
I want to be able to connect Apache Kafka to Apache Camel,
so that I can define my rules, routing and ETL pipeline to process and route data from and to different systems.5 votes -
CSV kafka connector
As application engg,
I want to develop a csv connector where data is ingested from flat files (CSV) and create a stream of records that can be processed in Apache Kafka. It is similar to what confluent kafka provides here (https://docs.confluent.io/kafka-connectors/spooldir/current/connectors/csv_source_connector.html)4 votes -
SAP Source connector for Aiven for Apache Kafka Connect
As a developer,
I want to be able to pull data from SAP systems,
so I could build my data pipelines.4 votes -
Kafka Connect GCS Sink: Support using field values to define bucket name or file name prefix
As data streaming architect
I want to be able to export records from Kafka to GCS and use values in the record to define the bucket or file name
so that I can organize data by those values to make them easier to find and process.Use case is a multi-user/multi-tenant application where user info is a value in the record. Need to be able to organize the output in object storage by that value somehow.
3 votes
- Don't see your idea?