101 results found
-
Schema subjects view with massive amounts of subjects
As a Avro schema administrator
I want to be able to quickly find an Avro schema subject in schema registry
so that I can view the schema, compare schema versions and change compatibility level.
In addition, we have 5-10k schema subjects in some of our projects, and there is no paging in the schema subject view - all subjects are loaded which makes the view very slow.Possible solutions might be to add paging, or simply limiting the number of subjects listed.
4 votes -
ClickHouse Driver for Apache Kafka Connect JDBC Sink connector
As a developer,
I want to have ClickHouse driver support in JDBC sink connector,
so that I can write data from Apache Kafka to ClickHouse for further processing and analysis.7 votes -
Include custom Kafka topic tags as Prometheus metric labels
The Aiven platform allows custom key/value tags to be added to resources such as Kafka topics. It would be useful to have these exposed as additional labels on metrics so that alerts can be triggered based on this metadata.
3 votes -
AWS SQS Source and Sink connector
As a developer,
I want to read and write data in AWS SQS
so that I can enable more architectural solutions and use SQS with other AWS services.11 votes -
Ability to scale Kafka cluster without upgrade
As a Kafka Administrator
I want to ahve ability to scalse Kafka clsuter without upgrade, that increses time of scale.
In production this can become critical and lead to downtime. For example last 2 times the scale lasted for 12 hours.1 vote -
A no-code solution for Flink to unlock the usage for non-technical users
As a non-technical user
I want to be able to aggregate and join different streams of data
without the need of developers1 voteThis function will largely be replaced by ChatGPT or other LLMs which can generate clear code and instructions, making a visual builder unnecessary.
-
Add Oracle JDBC support for Aiven for Kafka Connect
As a developer,
I want to be able to connect to my Oracle database from Aiven for Apache Kafka Connect,
so that I can read and write data from/to Oracle DB and/or Apache Kafka to enable my use-cases.9 votes -
Customer is interested to have Pyflink be supported with Flink
As a Developer I want to use Pyflink library with Aiven for Apache Flink so that I can directly use it in my project.
7 votes -
Tagging functionality for Apache Kafka ACLs
As a developer,
I want to tag Apache Kafka ACLs,
so that I can group, search and identify what my ACLs are used for.7 votes -
Support Zookeeper-less (KRaft) mode in Aiven for Apache Kafka
As a SRE,
I want to operate Apache Kafka without Zookeeper,
so that I can have more resources available for Apache Kafka itself.
In addition, it will allow faster up- and down- scaling of my cluster and it will support more partitions per broker.8 votes -
Add MaxTimestampSpec data to the topic/partition details from the Aiven API
As a developer,
I want to know latest offset and max timestamp for a given topic/partition,
so that I can identify when topic/partition was last time used and in case it is no longer in use I can cleanup stale topics/partitions.6 votes -
Support for cross-project and cross-vpc service integration for all services in streaming domain
As a developer,
I want to integrate my services in streaming domain, i.e. Apache Kafka, Apache Kafka Connect, Apache MirrorMaker2, etc with other services in any project and/or VPC,
so that I can build data streaming pipelines and benefit from the use use of aforementioned services.6 votes -
Group topics by type when displayed in Aiven Console
As a developer / operator,
I want to see Kafka topics grouped by type in Aiven Console
so that I can distinguish between topics created by Schema Registry, MirrorMaker2, Connect and directly in Kafka itself.
In addition, i want to apply filters and sorting by type.2 votes -
Aiven for Apache MirrorMaker2 should replicate ACLs
As a developer,
I want to replicate ACLs when migrating or duplicating my Apache Kafka clusters,
so that I can make sure my applications have the same rights and permissions on several clusters and I don't need to manually replicate them.6 votes -
Ease the setup of Kafka Connectors when using internal Aiven services
As a developer,
I want to have an easy setup of my Kafka connectors when they are using internal Aiven services such as Postgres or OpenSearch
so that I can save time, avoid mistakes, and enforce the strength of the platform.2 votes -
SAP Source connector for Aiven for Apache Kafka Connect
As a developer,
I want to be able to pull data from SAP systems,
so I could build my data pipelines.4 votes -
Flink HTTP API sink
As a developer I would like to push the output of a Flink operation to an HTTP API sink.
3 votes -
Improve support for Debezium SQL Server use cases
As an architect, I want to bring business logic out of the database and into a decoupled stream processing / event driven architecture framework. With SQL Server, I want to stream changes to Apache Kafka, using Debezium. This must support use cases where sensitivity classifications in SQL Server are used/required for things like PII.
2 votes -
Add external Delta Lake connector for Apache Flink
As a developer, I need to be able to read and write from Databricks so that I can complement my existing data architecture
2 votes -
Auto-Scaling
As a developer, I need Flink to be able to automatically increase compute resources to a self-defined threshold to seamlessly process significant increases in traffic in order to maintain performance. This increase in traffic would be for either a permanently higher workload, or for a temporary pre-defined amount of time, and would be triggered by a customer request, with subsequent additional billing impacts
2 votes
- Don't see your idea?