85 results found
-
Expose Karapace version
As a developer / user,
I want to know which Karapace version is currently running,
so that I can adapt new features and behaviours faster.
In addition, I would like to have version exposed in logs, HTTP responses and Aiven Console.8 votes -
Audit logging as self-service option
As a developer,
I want to enable audit logging for my service,
so that I can keep track of breadcrumbs left by all the changes made to a service to ensure compliance.8 votes -
Enforcing naming convention for Apache Kafka topics
As a SRE / Apache Kafka operator / developer,
I want to make sure all my topics adhere to a naming convention,
so that I can ensure consistency of my topics.
In addition, I can use naming convention to better identify, group, locate and categorise my topics.8 votes -
ClickHouse sink for Kafka Connect
As a developer,
I want to sync my data in to Clickhouse,
so that I can store large volumes of data and run analytics on top of it.8 votes -
Customer is interested to have Pyflink be supported with Flink
As a Developer I want to use Pyflink library with Aiven for Apache Flink so that I can directly use it in my project.
7 votesGathering Interest · AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded -
Kafka fined grained ACLs
As an ops engineer,
I want to be able to declare fined grained ACLs
so that I avoid having to grant "admin" rights to users which only need "DeleteRecords" rights on specific topics.
The client uses Kafka Streams, which needs specific rights (https://docs.confluent.io/platform/current/streams/developer-guide/security.html#required-acl-setting-for-secure-ak-clusters) which are not covered by aiven predefined rights.
Currently, the "admin" is to broad for such access (I don't want the user to be able to create topics).6 votes -
Add MaxTimestampSpec data to the topic/partition details from the Aiven API
As a developer,
I want to know latest offset and max timestamp for a given topic/partition,
so that I can identify when topic/partition was last time used and in case it is no longer in use I can cleanup stale topics/partitions.6 votes -
ClickHouse Driver for Apache Kafka Connect JDBC Sink connector
As a developer,
I want to have ClickHouse driver support in JDBC sink connector,
so that I can write data from Apache Kafka to ClickHouse for further processing and analysis.6 votes -
Support for cross-project and cross-vpc service integration for all services in streaming domain
As a developer,
I want to integrate my services in streaming domain, i.e. Apache Kafka, Apache Kafka Connect, Apache MirrorMaker2, etc with other services in any project and/or VPC,
so that I can build data streaming pipelines and benefit from the use use of aforementioned services.6 votes -
Ability to choose the Apache Kafka Connect connector version
As a developer,
I want to choose which Apache Kafka Connect connector version to use,
so that I can control connector version and make sure it is compatible with my applications.6 votes -
Tagging functionality for Apache Kafka ACLs
As a developer,
I want to tag Apache Kafka ACLs,
so that I can group, search and identify what my ACLs are used for.6 votes -
Azure Cosmos DB Kafka Connector
As a developer,
I want to connect to Azure Cosmos DB from Apache Kafka,
so that I can access my data stored in Cosmos DB and use it in my stream processing pipeline.6 votes -
Aiven for Apache MirrorMaker2 should replicate ACLs
As a developer,
I want to replicate ACLs when migrating or duplicating my Apache Kafka clusters,
so that I can make sure my applications have the same rights and permissions on several clusters and I don't need to manually replicate them.5 votes -
Support Zookeeper-less (KRaft) mode in Aiven for Apache Kafka
As a SRE,
I want to operate Apache Kafka without Zookeeper,
so that I can have more resources available for Apache Kafka itself.
In addition, it will allow faster up- and down- scaling of my cluster and it will support more partitions per broker.5 votes -
Apache Camel connector for Kafka Connect
As a software / data science engineer,
I want to be able to connect Apache Kafka to Apache Camel,
so that I can define my rules, routing and ETL pipeline to process and route data from and to different systems.5 votes -
Salesforce Kafka Connector
As a developer,
I want to connect my Apache Kafka to Salesforce
so that I can read and write data from/to SFDC.5 votes -
CSV kafka connector
As application engg,
I want to develop a csv connector where data is ingested from flat files (CSV) and create a stream of records that can be processed in Apache Kafka. It is similar to what confluent kafka provides here (https://docs.confluent.io/kafka-connectors/spooldir/current/connectors/csv_source_connector.html)4 votes -
Aiven's S3 sink connector - Support configure offset flush max size
As Developer
I want to use Aiven's S3 sink connector and have the ability to control the size of output files (keep same size even if the traffic changes). In general, the connector has no lags and we want to flush the offset and write to file only when we have enough data. We useoffset.flush.interval.ms
for it but when the traffic increases, the amount of data arrives in the configured interval is increased and can cause an OOM issue. In addition, when we pause the connector for couple of minutes and we gather a lag, it can also lead…4 votes -
Support for Protobuf serialization of Events (and Keys) in Aiven Kafka Connnect
As a data engineer
I want Aiven Kafka Connect to offer the option to use the protocol buffer data format when serializíng the events that is sends to a Kafka broker.
In my specific case, I need it to be possible in a Debezium connector for PostgreSQL.
Additionally, it would be good to have as a user the option to define oneself the protobuf schema to use for serializing.4 votes -
Karapace Schema Registry certificate authentication
As a developer I should be able to authenticate with same service user certificate for both Kafka and Schema Registry
4 votes
- Don't see your idea?