94 results found
-
Configure CPU threshold for alert mails in kafka
As a kafka user, I want to configure the threshold used to send CPU warning emails. Right now we are getting emails because the CPU exceeds 50%, which is irrelevant. We know we can configure the emails to go to another mail address but in that case we would loose all other technical mails as well.
It would also be good for us if we can turn off that single alert.2 votes -
MirrorMaker 2 connectivity for Kafka Brokers with private or internal hostnames
If an external Kafka service has brokers with hostnames that are not publicly resolvable, for example, where hostnames are registered in a corporate DNS server which is not accessible from outside, it is currently not possible for MM2 to successfully connect.
The workaround to use IP addresses is not sufficient, since after bootstrap the broker hostnames will be returned and subsequent connections will go there.
We will also likely see issues with TLS handshakes when brokers present certificates belonging to hostnames that the Aiven side does not understand.
The only current workaround is for the configuration of the external Kafka…
2 votes -
Decompression transform for Connectors
As an application developer
I want to compress my kafka messages, but be able to decomrpess them using a transform before sinking them into a destination
so that I can save on storage costs
In addition, I'd like to use ZTSD, but more common libraries might be enough.Note, Confluent has something similar :
https://docs.confluent.io/platform/current/connect/transforms/gzipdecompress.html1 vote -
Support S3 as a source and sink for Flink
As a developer I want to be able to read and write data to my S3 object storage, in order to simply integrate Flink into my existing data architecture. Using Flink to read data from S3, transform it, and then write to another S3 location allows easy consolidation and data quality management in a common reference data architecture.
1 vote -
Automated dynamic quota configuration
As a Cloud platform engineer, I need to have an automated way to set up and update quota configurations on a cluster, taking into account changes in resources consumption patterns amongst producers and consumers.
1 vote -
Ability to choose Apache Kafka service user password length using Aiven Console
As a developer,
I want to be able to select password length and difficulty when creating a service user,
so that I can stay complaint with internal security and compliance rules.2 votes -
Send a user alert in case stale credentials are being detected in connector configuration for Aiven for Apache Kafka Connect.
As a developer,
I want to be notified when I have stale credentials in my connectors,
so that I can update them and make sure connection from a connector to a source or sink system could be successfully established.2 votes -
Sending Apache Kafka metrics to Datadog
As Software Engineer at Wex
I want to send additional metrics to Datadog
so that I can send important metrics from Aiven's cluster as these ones:
IsrShrinksPerSec
IsrExpandsPerSec
ActiveControllerCount
OfflinePartitionsCount
TotalTimeMs
PurgatorySize
RequestsPerSec
Network bytes sent/received
BytesInPerSec/BytesOutPerSecAccording to Datadog's documentation, these metrics are considered highly significant.
https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/#kafka-emitted-metrics2 votes -
Debezium Connector for Cassandra
As a developer,
I want to connect to my Cassandra databases,
so that I can use Change Data Capture (CDC) pattern to be notified about data changes and can ingest that data in to my data streaming pipeline for further use and processing.2 votes -
io.debezium.transforms.partitions.PartitionRouting
It would be great it the
io.debezium.transforms.partitions.PartitionRouting
SMT was available for use when configuring a Kafka Debezium source connector on the Aiven platform.1 vote -
Expose MirrorMaker 2 replication.policy.separator property
As a devops developer
I want to change the replication policy separator
so that I can use our existing topic separators without conflicting with MM2.
In addition, I find it extremely inconvenient to have to change our existing schema because we cannot configure the replication.policy.separator in MM2.1 voteWe can see it as useful configuration to expose. At the same time it will need to gather a bit more interest to be taken in to development.
Thanks!
-
Support for schemaless JSON messages in the Big Query Sink connector
As a developer
I want to be able to publish schemaless JSON messages to Big Query and have the BigQuery schema be updated to reflect those changes
so that I can evolve my message schema without breaking my pipeline.1 vote -
Rockset Sink connector for Kafka Connect
As a data engineer,
I want to write data to Rockset,
so that I can run my analytical workloads.1 vote -
Apache NiFi Source and Sink Kafka connector
As a developer
I want to use Apache NiFi Kafka Connector
so that I can move data in and out Apache NiFi to orchestrate data flow.1 vote
- Don't see your idea?