91 results found
-
Azure Blob Storage Kafka connector
As a developer,
I want to read and write data to Azure Blob Storage,
so that I can use that data for stream processing and analytics, and backup / restore data in Apache Kafka.16 votesRoadmapped ·AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded
-
Apache Camel connector for Kafka Connect
As a software / data science engineer,
I want to be able to connect Apache Kafka to Apache Camel,
so that I can define my rules, routing and ETL pipeline to process and route data from and to different systems.5 votes -
ClickHouse sink for Kafka Connect
As a developer,
I want to sync my data in to Clickhouse,
so that I can store large volumes of data and run analytics on top of it.10 votes -
Salesforce Kafka Connector
As a developer,
I want to connect my Apache Kafka to Salesforce
so that I can read and write data from/to SFDC.6 votes -
Debezium Connector for Cassandra
As a developer,
I want to connect to my Cassandra databases,
so that I can use Change Data Capture (CDC) pattern to be notified about data changes and can ingest that data in to my data streaming pipeline for further use and processing.2 votes -
Rockset Sink connector for Kafka Connect
As a data engineer,
I want to write data to Rockset,
so that I can run my analytical workloads.1 vote -
Azure Cosmos DB Kafka Connector
As a developer,
I want to connect to Azure Cosmos DB from Apache Kafka,
so that I can access my data stored in Cosmos DB and use it in my stream processing pipeline.6 votes -
Support for AWS S3 Source Kafka Connector
As a developer / data engineer,
I want to be able to read data stored in AWS S3 bucket,
so that I can transfer, process and transform that data to other applications.
In addition, I can use data stored in S3 bucket as backup and I can rehydrate my Apache Kafka with it.16 votesRoadmapped ·AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded
-
AWS SQS Source and Sink connector
As a developer,
I want to read and write data in AWS SQS
so that I can enable more architectural solutions and use SQS with other AWS services.10 votes -
Secure secrets in Hashicorp Vault and use them in Kafka Connect.
As DevOps/SecOps engineer,
I want to securely manage my credentials in Kafka Connect connectors,
so that I can make sure no credentials are being leaked in connector configuration.
In addition, I can manage my credentials in Hashicorp Vault and exercise credential rotation without changing my connector configuration.26 votes -
Apache NiFi Source and Sink Kafka connector
As a developer
I want to use Apache NiFi Kafka Connector
so that I can move data in and out Apache NiFi to orchestrate data flow.1 vote
- Don't see your idea?