7 results found
-
Support exactly-once semantics when connecting Kafka to ClickHouse
As a developer using Kafka to stream messages to ClickHouse,
I want to be sure that messages as delivered and ingested exactly once, as opposed to at-least-once, to ClickHouse to not have to deal with duplicated data that I have to deduplicate down the line. This removes both a cognitive and operational load from my data pipeline
c.f. https://github.com/ClickHouse/clickhouse-kafka-connectThe exactly-once semantics has been implemented in the newer version of the ClickHouse Kafka Connect Sink (https://github.com/ClickHouse/clickhouse-kafka-connect).
5 votes -
MongoDB functions and table engine.
As someone running analytics with Clickhouse,
I want to be able to create functions that read data from an external MongoDB instance, as well as create tables that span data in external MongoDB instances,
so that I can enrich my analytics with data stored elsewhere, or easily migrate data from a MongoDB instance to my Aiven for Clickhouse instance.5 votes -
Option for ClickHouse Tiered Storage to use my own object storage
As a user of Aiven for ClickHouse on AWS, Microsoft Azure or Google Cloud:
I would like to leverage my own object storage account with Tiered Storage for Aiven for ClickHouse (that I am already using BYOC on).4 votes -
Add support for the DeltaLake Engine to ClickHouse
As a customer of Aiven, I would like to pull data from many disparate sources into my ClickHouse data warehouse to provide my users best-in-class analytics and performance. Adding support for DeltaLake would open up new sources from which I can seamlessly consume from.
4 votes -
Support for Postgresql table engine
As a developer with a current PostgreSQL instance,
I would like to utilise the ClickHouse PostgreSQL Table engine,
so that I can easily read from that external PostgreSQL instance and insert the data into my Aiven for ClickHouse instance.3 votes -
MySQL Table engine
As someone drunning analytics queries with Clickhouse against multiple data sources,
I want to have the MySQL table engine enabled so that I can create tables in Clickhouse that span data in external MySQL instances,
so that I can bring the value of external data sources into my analysis.2 votes -
Manage ClickHouse integrations' credentials securely & easily
As a ClickHouse user, I want to be able to use table functions & engines that require identification credentials (ex. private remote s3 bucket, remote delta lake table etc.) without hassle and risks.
With named collections in Aiven for ClickHouse, you can set your integrations credentials once and use it safely with all your remote queries.
Moreover, you can easily rotate credentials if needed: change credentials once using the Aiven console and apply it to all your integrations.1 vote
- Don't see your idea?