Skip to content

Data Analytics

Join our forum to discuss your ideas with Aiven community or check out our public roadmap.

Data Analytics

Categories

JUMP TO ANOTHER FORUM

34 results found

  1. I need to read and write Parquet format files to/from Azure Blob Storage.

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  2. As a user sending Kafka messages to Aiven for ClickHouse,
    I want to be able to send those messages using the protobuf format

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  3. As a user of Aiven for ClickHouse on AWS, Microsoft Azure or Google Cloud:
    I would like to leverage my own object storage account with Tiered Storage for Aiven for ClickHouse (that I am already using BYOC on).

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  4. As a developer using Kafka to stream messages to ClickHouse,
    I want to be sure that messages as delivered and ingested exactly once, as opposed to at-least-once, to ClickHouse to not have to deal with duplicated data that I have to deduplicate down the line. This removes both a cognitive and operational load from my data pipeline
    c.f. https://github.com/ClickHouse/clickhouse-kafka-connect

    The exactly-once semantics has been implemented in the newer version of the ClickHouse Kafka Connect Sink (https://github.com/ClickHouse/clickhouse-kafka-connect).

    4 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  5. As a user of both PostgreSQL (on-prem, on another cloud provider or at Aiven) and Aiven for ClickHouse,
    I want to be able to ingest my PostgreSQL data, not as a on time snapshot or a remote view of the data like the current integrations allow me to, but as tables created and stored in ClickHouse that pull updates from PostgreSQL regularly.

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  6. As developer against ClickHouse,
    I want to utilise the S3Queue engine available in v23.8
    so that I can watch an S3 bucket, and auto consume new files as they appear, via a materialized view, as soon as they arrive.

    This will enable me to keep my analytics up to date with no need for additional engineering effort or streaming pipelines.

    6 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  7. As a developer, as a DR requirement, I would like to be able to restore the service from its backup in cases where the service's hosting region is down.

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  8. As a business owner,
    I would like to have the ability to set up a replication of my Aiven for ClickHouse service, and all the object storage attached to it, to a secondary region,
    so that in the case of needing disaster recovery from the primary region going down, we could fail our workload traffic over to the secondary region.

    The replicated secondary region does not need to be writable until the primary is unavailable, but we would like to read from it to validate consistency. As a customer we'll handle the failover mechanism.

    7 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  9. At this moment there are only the primary Postgres instances available for a clickhouse connect. Wouldn't it make sense to have the capability connection Clickhouse to a read replica to prevent performance impacts on the primary?

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  10. As a developer with a current PostgreSQL instance,
    I would like to utilise the ClickHouse PostgreSQL Table engine,
    so that I can easily read from that external PostgreSQL instance and insert the data into my Aiven for ClickHouse instance.

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  11. As an Aiven for ClickHouse user,
    I want to be able to use the query editor in the Aiven console in my ClickHouse services to run large queries without worrying of timeouts or bugs.

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  12. As a user of Aiven for ClickHouse,
    I want to use the most optimized default settings possible. From the start I want to have the default JOIN set as the most performant possible and not the ClickHouse default one. This comes without me, the user, having to change any settings. However, I should be able to tweak those settings if needed in specific cases.
    See https://clickhouse.com/docs/en/engines/table-engines/special/join

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  13. As a user of Aiven for ClickHouse,
    I want to be able to choose between Aiven's object storage (the default) and my own Oracle - OCI object storage.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  14. As a customer of Aiven, I would like to pull data from many disparate sources into my ClickHouse data warehouse to provide my users best-in-class analytics and performance. Adding support for DeltaLake would open up new sources from which I can seamlessly consume from.

    4 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  15. As a user,
    I want to have a better idea why queries I run in the built-in query editor in the Aiven console do not work.
    Clearer error messages are a way to let users know what went wrong and modify their queries accordingly.

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  16. As someone managing a Clickhouse instance that delivers my production workloads,
    I want to have better insight into what is causing slow queries against my data,
    so that I can get to the root cause and take corrective actions.
    In addition, it's important to have these metrics available for export to my monitoring system

    6 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  17. As a user of Aiven for ClickHouse,
    I want to be able to know more about Zookeeper, its current state and what it is doing at any given time (in the console or in Grafana).

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  18. As a user of at least one Aiven for ClickHouse service,

    I want to be able to easily view in the Aiven console, in my ClickHouse service, the list of mutations (both in-progress and completed) and relevant information about them such as creation time, ID of the mutation, what table and databases the mutation is applied on, if it is done or in progress etc.

    At the moment, this information can be found in system.mutations table and system.replication_queue for replicated tables, but it is a manual process to check them.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  19. As developer ina. company with a large amount of data stored in Iceberg,
    I want to have the ClickHouse Iceberg Table engine,
    so that I can effectively analyse the large amounts of data stored in iceberg, via Aiven for Clickhouse.

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  20. As a user of Apache Beam and ClickHouse,
    I want to be able to ingest my data from Apache Beam into a ClickHouse table in near-real time with at-least-once (meaning it is acceptable that messages are read more than once). Writing the data using the ClickHouseIO (built on JDBC driver) c.f. https://beam.apache.org/releases/javadoc/2.10.0/org/apache/beam/sdk/io/clickhouse/ClickHouseIO.html

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
← Previous 1
  • Don't see your idea?