91 results found
-
Programmatic Apache Kafka Consumer (group) management
As a developer,
I want to programmatically manage my consumers and consumer groups,
so that I can see the status of them, perform CRUD operations, show members of a group, reset offsets and similar.
In addition, I want to be able to do so also in Aiven Console.18 votes -
Support Databricks driver for Apache Kafka Connect JDBC sink connector
As a developer,
I want to have support for Databricks driver for Apache Kafka Connect JDBC sink connector,
so that I can write data from Aiven for Apache Kafka to Databricks Spark for further processing, analysis and consumption.11 votes -
ClickHouse Driver for Apache Kafka Connect JDBC Sink connector
As a developer,
I want to have ClickHouse driver support in JDBC sink connector,
so that I can write data from Apache Kafka to ClickHouse for further processing and analysis.7 votes -
Audit logging as self-service option
As a developer,
I want to enable audit logging for my service,
so that I can keep track of breadcrumbs left by all the changes made to a service to ensure compliance.11 votes -
Configurable maximum validity for service user certificates
As a developer / security engineer,
I want to be able to define lifetime of certificates (i.e. 30-90 days) used to authenticate my service users,
so that I can have certificate rotation policies in place to ensure compliance and security best practices.10 votes -
Enforcing naming convention for Apache Kafka topics
As a SRE / Apache Kafka operator / developer,
I want to make sure all my topics adhere to a naming convention,
so that I can ensure consistency of my topics.
In addition, I can use naming convention to better identify, group, locate and categorise my topics.9 votes -
Support for cross-project and cross-vpc service integration for all services in streaming domain
As a developer,
I want to integrate my services in streaming domain, i.e. Apache Kafka, Apache Kafka Connect, Apache MirrorMaker2, etc with other services in any project and/or VPC,
so that I can build data streaming pipelines and benefit from the use use of aforementioned services.6 votes -
Ability to choose the Apache Kafka Connect connector version
As a developer,
I want to choose which Apache Kafka Connect connector version to use,
so that I can control connector version and make sure it is compatible with my applications.6 votes -
Run Karapace as dedicated managed service
As a developer,
I want to run Karapace as fully managed dedicated service,
so that I can use it with Apache Kafka running on and outside of Aiven.
In addition, I can use same Karapace service against multiple Apache Kafka services.17 votes -
Aiven for Apache MirrorMaker2 should replicate ACLs
As a developer,
I want to replicate ACLs when migrating or duplicating my Apache Kafka clusters,
so that I can make sure my applications have the same rights and permissions on several clusters and I don't need to manually replicate them.5 votes -
Temporarily disable service user on Aiven for Apache Kafka
As a developer,
I want to temporarily disable service user,
so that I can test is service user still in use by any of the applications or temporarily ban abusive clients.10 votes -
Support for Apache Kafka Connect to be run standalone against a 3rd party Apache Kafka service
As a developer,
I want to run standalone Apache Kafka Connect cluster against 3rd party Apache Kafka service,
so that I can benefit from managed Apache Kafka Connect service and read/write data from/to Apache Kafka service running outside Aiven.10 votes -
Capture and expose "last used" date for service users
As a developer / SRE,
I want to know when service users were last used,
so that I can know is service user still in use or was abandoned and could be removed.10 votes -
Tagging functionality for Apache Kafka ACLs
As a developer,
I want to tag Apache Kafka ACLs,
so that I can group, search and identify what my ACLs are used for.6 votes -
Schema Validation on Apache Kafka broker side
As a developer,
I want to make sure schema is being validated not just on client but also on a broker side,
so that I can make sure all messages in the topic correspond to the same schema and does not contain any mixed schemas.18 votes -
Support for public CA for SASL for Aiven for Apache Kafka
As developer,
I want to use public CA (i.e. Let's Encrypt),
so that I can connect to my Apache for Kafka service without installing any additional certificates.
In addition, I can trust publish authority issuing certificates instead of validating 3rd party certificates.15 votes -
Support for schemaless JSON messages in the Big Query Sink connector
As a developer
I want to be able to publish schemaless JSON messages to Big Query and have the BigQuery schema be updated to reflect those changes
so that I can evolve my message schema without breaking my pipeline.1 vote -
Customer is interested to have Pyflink be supported with Flink
As a Developer I want to use Pyflink library with Aiven for Apache Flink so that I can directly use it in my project.
7 votesGathering Interest ·AdminMichael Tansini (Product Manager Data Streaming (Kafka and Flink), Aiven.io) responded
-
Support Zookeeper-less (KRaft) mode in Aiven for Apache Kafka
As a SRE,
I want to operate Apache Kafka without Zookeeper,
so that I can have more resources available for Apache Kafka itself.
In addition, it will allow faster up- and down- scaling of my cluster and it will support more partitions per broker.8 votes -
Allow inspected custom code (JARs) in Apache Flink and Apache Kafka Connect
As a developer,
I want to be able to upload my custom code (JARs),
so that I can use it to build sophisticated or non-standard use-cases in Flink and Kafka Connect,
In addition, I can keep custom implementation private.50 votes
- Don't see your idea?