FAQ

Find quick answers to common questions about Firebolt
How can query performance be optimized when querying event data with minute-level granularity in Firebolt?

One approach is to restructure the table by setting the primary index on event_time to better leverage Firebolt’s indexing capabilities. Additionally, an aggregating index on event_time can be beneficial. However, if queries still take longer than expected (e.g., 15 seconds for 30 days of data), it may help to review: - The structure of the primary index and ensure it aligns with the query’s filtering. - Whether unnecessary dimensions are included in the dataset, increasing granularity unnecessarily. - If joins or aggregations can be optimized, possibly through pre-aggregated tables. Firebolt’s architecture is designed to improve query efficiency by avoiding costly full scans and optimizing indexing structures.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Will the creation of an engine automatically result in the creation of the underlying cluster(s)?

Yes. By default, creating an engine would result in the creation of the underlying engine clusters and start the engine. This would enable the engine to be in a running state where it is ready to start serving the queries. However, you have the option to defer the creation of the underlying clusters for an engine by setting the property “INITIALLY STOPPED” to True while calling CREATE ENGINE. You can start the engine at a later point, when you are ready to start running queries on the engine. Note that you cannot modify this property after an engine has been created.

CREATE ENGINE IF NOT EXISTS MyEngine WITH

TYPE = “S” NODES = 2 CLUSTERS =1 START_IMMEDIATELY = FALSE;

Engines
COPY LINK TO ANSWER
will-the-creation-of-an-engine-automatically-result-in-the-creation-of-the-underlying-cluster-s

https://firebolt.io/faqs-v2-knowledge-center/will-the-creation-of-an-engine-automatically-result-in-the-creation-of-the-underlying-cluster-s

Can Firebolt handle complex data types like JSON during ELT processes?

Yes, Firebolt supports semi-structured data types like JSON. JSON data can be ingested as text columns or parsed into individual columns for flexible schema-on-read or flattened structures.

ELT
COPY LINK TO ANSWER
can-firebolt-handle-complex-data-types-like-json-during-elt-processes

https://firebolt.io/faqs-v2-knowledge-center/can-firebolt-handle-complex-data-types-like-json-during-elt-processes

What versions of Presto does Firebolt support?

Firebolt does not natively support Presto. However, Firebolt provides its own high-performance SQL engine and you can integrate it with your existing data infrastructure via ODBC, JDBC, and REST API.

Integrations
COPY LINK TO ANSWER
what-versions-of-presto-does-firebolt-support

https://firebolt.io/faqs-v2-knowledge-center/what-versions-of-presto-does-firebolt-support

How do you protect against DDoS attacks?

We use AWS Shield, WAF, and other logical layers to protect against DDoS. Additionally, we leverage auto-scaling to maintain availability during attacks by dynamically adjusting resources like EC2 instances, ELBs, and other global services capacity. (Though some scenarios may require manual intervention).

Security
COPY LINK TO ANSWER
how-do-you-protect-against-ddos-attacks

https://firebolt.io/faqs-v2-knowledge-center/how-do-you-protect-against-ddos-attacks

Can I query the output of another query in Firebolt?

1. Use a WITH clause to define a Common Table Expression (CTE) and query its result.

2. Create a VIEW based on the result set and query the VIEW.

SQL
COPY LINK TO ANSWER
can-i-query-the-output-of-another-query-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/can-i-query-the-output-of-another-query-in-firebolt

Is Firebolt billing real-time?

FBU consumption is reported in real time and can be used to calculate costs by multiplying the consumed FBU by the price listed on the pricing page or a custom deal rate. However, the Billing and Consumption page updates daily, and AWS storage costs have a ~48-hour delay. For more information, check our documentation.

Pricing & Billing
COPY LINK TO ANSWER
is-firebolt-billing-real-time

https://firebolt.io/faqs-v2-knowledge-center/is-firebolt-billing-real-time

How do I start and stop engines in Firebolt?

To start an engine:

sql
START ENGINE MyEngine;

To stop an engine:

vbnet
STOP ENGINE MyEngine;

For more information, please refer to the Work with Engines Using DDL article in the Firebolt Documentation.

Engines
COPY LINK TO ANSWER
how-do-i-start-and-stop-engines-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-start-and-stop-engines-in-firebolt

How do I handle complex structures like arrays in Parquet?

Firebolt supports the ARRAY data type for mapping arrays from Parquet files and provides functions like ARRAY_AGG and UNNEST for working with arrays.

ELT
COPY LINK TO ANSWER
how-do-i-handle-complex-structures-like-arrays-in-parquet

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-handle-complex-structures-like-arrays-in-parquet

How are SQL errors handled in the SDK?

Firebolt's Python SDK provides detailed error message handling for SQL queries. When an error occurs, the SDK generates helpful error messages, allowing users to quickly diagnose and fix issues such as syntax problems or missing credentials. The SDK also offers robust logging and debugging capabilities, making it easier for developers to troubleshoot errors in their applications. For more information, refer to the Firebolt Python SDK documentation or visit the GitHub repository for examples.

Integrations
COPY LINK TO ANSWER
how-are-sql-errors-handled-in-the-sdk

https://firebolt.io/faqs-v2-knowledge-center/how-are-sql-errors-handled-in-the-sdk

How does Firebolt protect customer data?

Firebolt employs a comprehensive security strategy that includes network security policies, encryption practices, tenant isolation, and governance controls. We are committed to safeguarding your data through state-of-the-art security systems, policies, and practices.

Security
COPY LINK TO ANSWER
how-does-firebolt-protect-customer-data

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-protect-customer-data

How can I make sure that my engines are not sitting idle and incurring infrastructure costs?

You can use the AUTO_STOP feature available in Firebolt engines to make sure that your engines are automatically stopped after a certain amount of idle time. Engines in stopped state will not be charged, hence do not incur any costs. As with other engine operations, this can be done via SQL or the UI. For example, while creating an engine, you can specify the idle time, using AUTO_STOP, as below:

CREATE ENGINE IF NOT EXISTS MyEngine WITH
TYPE = “S” NODES = 2 CLUSTERS =1 AUTO_STOP = 15;

The above command will ensure that MyEngine will be automatically stopped if it has been idle for 15 minutes continuously. Alternatively, you can achieve the same after an engine has been created.

ALTER ENGINE MyEngine SET AUTO_STOP = 15;

For mor information, please see the Engine Consumption Documentation.

Engines
COPY LINK TO ANSWER
how-can-i-make-sure-that-my-engines-are-not-sitting-idle-and-incurring-infrastructure-costs

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-make-sure-that-my-engines-are-not-sitting-idle-and-incurring-infrastructure-costs

What is multistage distributed execution in Firebolt, and how does it improve ELT operations?

Multistage distributed execution allows complex ELT queries to utilize all cluster resources by splitting stages across nodes. This parallelization optimizes resource utilization, speeding up ELT processes.

ELT
COPY LINK TO ANSWER
what-is-multistage-distributed-execution-in-firebolt-and-how-does-it-improve-elt-operations

https://firebolt.io/faqs-v2-knowledge-center/what-is-multistage-distributed-execution-in-firebolt-and-how-does-it-improve-elt-operations

How do I handle access errors when using service accounts in Firebolt?

When using service accounts in Firebolt, access errors can occur due to incorrect credentials, missing permissions, or expired tokens. To troubleshoot these errors, follow these steps:

- Check your credentials: Ensure that the service account credentials (client ID and secret) are correct and active.
- Verify permissions: Make sure the service account has the appropriate role or permissions to access the resources you are trying to interact with, such as S3 buckets or Firebolt tables.
- Refresh authentication tokens: If you're using tokens, ensure they have not expired. Tokens generated for service accounts typically have a limited lifespan.
- Logs and error details: Review the error message logs for more specific information about the access failure.
- Review documentation: Follow the Firebolt documentation on service accounts for proper setup, permissions, and token management.

Integrations
COPY LINK TO ANSWER
how-do-i-handle-access-errors-when-using-service-accounts-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-handle-access-errors-when-using-service-accounts-in-firebolt

Is IP Allow-listing supported for tenant connections?

Yes, both IP Allow/Deny-listing is supported. More details on our Network Policy page.

Security
COPY LINK TO ANSWER
is-ip-allow-listing-supported-for-tenant-connections

https://firebolt.io/faqs-v2-knowledge-center/is-ip-allow-listing-supported-for-tenant-connections

What happens when an engine receives queries while it is in a stopped state?

If the engine has the AUTO_START option set to True, an engine in a stopped state will be automatically started when it receives a query. By default, this option is set to True. If this option is set to False, you must explicitly start the engine using the START ENGINE command. For more information, please refer to the Work with Engines Using DDL article in the Firebolt Documentation.

Engines
COPY LINK TO ANSWER
what-happens-when-an-engine-receives-queries-while-it-is-in-a-stopped-state

https://firebolt.io/faqs-v2-knowledge-center/what-happens-when-an-engine-receives-queries-while-it-is-in-a-stopped-state

What options does Firebolt provide to export data?

Use the COPY TO SQL command to export data to S3.

ELT
COPY LINK TO ANSWER
what-options-does-firebolt-provide-to-export-data

https://firebolt.io/faqs-v2-knowledge-center/what-options-does-firebolt-provide-to-export-data

How do I work with authentication tokens in Firebolt?

To work with authentication tokens in Firebolt:

Generate a token via Firebolt’s authentication endpoint using your client ID and secret.

Example:

curl -X POST https://id.app.firebolt.io/oauth/token \
--header 'Content-Type: application/x-www-form-urlencoded' \
--data-urlencode 'grant_type=client_credentials' \
--data-urlencode 'client_id=YOUR_CLIENT_ID' \
--data-urlencode 'client_secret=YOUR_CLIENT_SECRET'

Use the token in API requests by including it in the authorization header:

--header 'Authorization: Bearer YOUR_ACCESS_TOKEN'

Refresh tokens regularly, as they expire. Keep tokens secure using environment variables or secret managers.

For more details, refer to the Firebolt API documentation.

Integrations
COPY LINK TO ANSWER
how-do-i-work-with-authentication-tokens-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-work-with-authentication-tokens-in-firebolt

Can customers federate users to their chosen identity provider (IdP)?

Yes, this feature is supported. More details on our Identity Management page.

Security
COPY LINK TO ANSWER
can-customers-federate-users-to-their-chosen-identity-provider-idp

https://firebolt.io/faqs-v2-knowledge-center/can-customers-federate-users-to-their-chosen-identity-provider-idp

How do I monitor the performance of my engine to understand whether it is optimally configured?

Firebolt provides three different observability views that provide insight into the performance of your engine.

1/ engine_running_queries - This view provides Information about currently running queries. This includes whether a query is running or in the queue. For queries that are currently running, this view also provides information on how long it has been running.

2/ engine_query_history - This view provides historical information about past queries - for each query in history, this includes the execution time of the query, amount of CPU and Memory consumed and amount of time the query spent in queue, among other details.

3/ engine_metrics_history - This view provides information about the utilization of CPU, RAM and Storage for each of the engine clusters. You can use these views to understand whether your engine resources are being utilized optimally, whether your query performance is meeting your needs,  what percentage of queries are waiting in the queue and for how long. Based on these insights, you can resize your engine accordingly.

For more information, please refer to the Sizing Engines article in our Documentation.

Engines
COPY LINK TO ANSWER
how-do-i-monitor-the-performance-of-my-engine-to-understand-whether-it-is-optimally-configured

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-monitor-the-performance-of-my-engine-to-understand-whether-it-is-optimally-configured

What file formats are supported by 'COPY TO'?

CSV, TSV, JSON, and Parquet formats are supported for exporting data.

ELT
COPY LINK TO ANSWER
what-file-formats-are-supported-by-copy-to

https://firebolt.io/faqs-v2-knowledge-center/what-file-formats-are-supported-by-copy-to

Is Multi-Factor Authentication (MFA) supported?

Yes, MFA is supported. More details on our MFA page.

Security
COPY LINK TO ANSWER
is-multi-factor-authentication-mfa-supported

https://firebolt.io/faqs-v2-knowledge-center/is-multi-factor-authentication-mfa-supported

How can I monitor the consumption of engines?

Use the engine_metering_history information schema view to track FBU consumption for each engine.

Engines
COPY LINK TO ANSWER
how-can-i-monitor-the-consumption-of-engines

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-monitor-the-consumption-of-engines

Can I automate ELT processes with Firebolt?

Yes, ELT processes can be automated using: Firebolt Python SDK for programmatic database operations. Apache Airflow for scheduling and automating complex workflows. dbt (Data Build Tool) for managing data transformations in a version-controlled environment.

ELT
COPY LINK TO ANSWER
can-i-automate-elt-processes-with-firebolt

https://firebolt.io/faqs-v2-knowledge-center/can-i-automate-elt-processes-with-firebolt

Does Firebolt support RBAC?

Yes, RBAC is supported. More details on our RBAC page.

Security
COPY LINK TO ANSWER
yes-rbac-is-supported-more-details-on-our-rbac-page

https://firebolt.io/faqs-v2-knowledge-center/yes-rbac-is-supported-more-details-on-our-rbac-page

Can users see what is currently running on a specific engine when it is at 100% utilization?

Yes. The engine_running_querie and engine_query_history tables provide insights into current workloads. For more information see our Information Schema Documenation.

Engines
COPY LINK TO ANSWER
can-users-see-what-is-currently-running-on-a-specific-engine-when-it-is-at-100-utilization

https://firebolt.io/faqs-v2-knowledge-center/can-users-see-what-is-currently-running-on-a-specific-engine-when-it-is-at-100-utilization

Our ELT jobs are expensive and impact customer-facing dashboards. How can Firebolt address this?

Firebolt enables running ELT jobs on a separate engine isolated from the customer-facing engine. This prevents disruptions and allows scaling ELT engines dynamically with auto-start and auto-stop features to reduce costs.

ELT
COPY LINK TO ANSWER
our-elt-jobs-are-expensive-and-impact-customer-facing-dashboards-how-can-firebolt-address-this

https://firebolt.io/faqs-v2-knowledge-center/our-elt-jobs-are-expensive-and-impact-customer-facing-dashboards-how-can-firebolt-address-this

Is data sovereignty supported?

Yes, customers can choose the region in which they run the service to meet data sovereignty requirements. More on our Available Region page.

Security
COPY LINK TO ANSWER
is-data-sovereignty-supported

https://firebolt.io/faqs-v2-knowledge-center/is-data-sovereignty-supported

How can I check Firebolt engine status via the REST API?

Monitoring the status of your Firebolt engine using the REST API is a key step to ensure smooth operations. Firebolt provides a way to programmatically check engine status by querying the system engine. This article explains how to authenticate, retrieve the system engine URL, and query the engine status using Firebolt's REST API.To begin, ensure that you have an active service account with the necessary permissions. You will need the service account credentials to generate an access token for API authentication.

After obtaining an access token, use the following request to retrieve the system engine URL:

curl https://api.app.firebolt.io/web/v3/account/<account_name>/engineUrl \
-H 'Accept: application/json' \
-H 'Authorization: Bearer <access_token>'

Once you have the system engine URL, you can query it to check the engine's status with a simple SQL query, as shown below:

curl --location 'https://<system_engine_URL>/query' \
--header 'Authorization: Bearer <access_token>' \
--data "select status from information_schema.engines where engine_name = '<your_engine_name>'

This will return the current status of your engine, helping you monitor its activity and health.

For more information please refer to the Using the API Documentation.

Engines
COPY LINK TO ANSWER
how-can-i-check-firebolt-engine-status-via-the-rest-api

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-check-firebolt-engine-status-via-the-rest-api

Can Firebolt support adding new columns to existing tables without rebuilding the entire table?

No, Firebolt doesn't support adding new columns without rebuilding the table. However, you can create a new table with the updated schema or use views to simulate schema changes.

This ensures that the schema remains optimized for performance, which is critical in high-performance analytical databases like Firebolt.

Alternatively, Firebolt offers a flexible approach where you can create views to simulate changes like renaming or restructuring tables without needing to rebuild or re-ingest data. For instance, you can create a view that selects all columns from the original table, effectively simulating the addition of new columns:

Example Usage: To simulate renaming a table or altering its structure, create a view:

CREATE VIEW IF NOT EXISTS new_games ASSELECT * FROM games;

This approach allows you to redirect queries to the new view (new_games), making it function like a table with updated schema without altering the original table.

ELT
COPY LINK TO ANSWER
can-firebolt-support-adding-new-columns-to-existing-tables-without-rebuilding-the-entire-table

https://firebolt.io/faqs-v2-knowledge-center/can-firebolt-support-adding-new-columns-to-existing-tables-without-rebuilding-the-entire-table

Is encryption supported for data at rest and in motion?

Yes, we support encryption for data at rest and in motion. More on our technical best practices can be found in our Security blog: “Building Customer Trust: A CISO's Perspective on Security and Privacy at Firebolt”

Security
COPY LINK TO ANSWER
is-encryption-supported-for-data-at-rest-and-in-motion

https://firebolt.io/faqs-v2-knowledge-center/is-encryption-supported-for-data-at-rest-and-in-motion

How do I calculate the consumption of engines?

Firebolt uses Firebolt Units (FBU) to track engine consumption. For example, for an engine with Type "S", 2 nodes, and 1 cluster running for 30 minutes, it would consume 8 FBUs:

FBU per Hour = 8 * 2 * 1 = 16 FBUs
Consumption = (16 / 3600) * 1800 seconds = 8 FBUs

Engines
COPY LINK TO ANSWER
how-do-i-calculate-the-consumption-of-engines

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-calculate-the-consumption-of-engines

How does Firebolt address complex queries that exceed physical memory capacity?

Firebolt distributes data across nodes and uses spilling to local SSDs when the working set exceeds available memory, allowing the system to scale even with limited resources.

ELT
COPY LINK TO ANSWER
how-does-firebolt-address-complex-queries-that-exceed-physical-memory-capacity

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-address-complex-queries-that-exceed-physical-memory-capacity

How does Firebolt mitigate SQL Injection attacks?

Firebolt database itself inherently reduces the risk of SQL injection by minimizing the use of certain vulnerable constructs. Customers are still encouraged to implement additional controls at their application level such as:
- Ensure all user inputs are strictly validated before being processed.
- Escape potentially dangerous characters that could be used in unexpected ways.
- Include SQL injection tests in your regular security testing and code review processes.

Security
COPY LINK TO ANSWER
how-does-firebolt-mitigate-sql-injection-attacks

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-mitigate-sql-injection-attacks

How can I make sure that my engines are not sitting idle and incurring infrastructure costs?

Use the AUTO_STOP feature to automatically stop engines after a certain amount of idle time. Example:

:ALTER ENGINE MyEngine SET AUTO_STOP = 15;

For more information, read more about Engine Consumption in our Documentation.

Engines
COPY LINK TO ANSWER
how-can-i-make-sure-that-my-engines-are-not-sitting-idle-and-incurring-infrastructure-costs-2

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-make-sure-that-my-engines-are-not-sitting-idle-and-incurring-infrastructure-costs-2

How do I troubleshoot errors when ingesting CSV files into Firebolt using external tables?

CSV file ingestion into an external table may fail with errors such as:

Cannot parse input: expected '|' but found '<CARRIAGE RETURN>' instead.

This usually means the file delimiter in the CSV doesn't match the table definition or the number of columns differs.

To troubleshoot check the delimiter: Ensure FIELD_DELIMITER matches the CSV file's delimiter. Also, compare the file to the table definition column-by-column. Finally, if the file is large, create a temporary external table to view the entire row as a single string:

CREATE EXTERNAL TABLE ext_tmp (blob text) URL = 's3://some_bucket/somefolder/' TYPE = (CSV FIELD_DELIMITER=' ');

Example Query:

SELECT * FROM ext_tmp LIMIT 2;

This helps inspect rows and verify column consistency.

ELT
COPY LINK TO ANSWER
how-do-i-troubleshoot-errors-when-ingesting-csv-files-into-firebolt-using-external-tables

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-troubleshoot-errors-when-ingesting-csv-files-into-firebolt-using-external-tables

How does Firebolt protect against malware, zero-day exploits, and other runtime threats?

Besides our runtime binary hardening, Firebolt leverages a runtime protection tool that provides deep visibility and protection at the process level.

Security
COPY LINK TO ANSWER
how-does-firebolt-protect-against-malware-zero-day-exploits-and-other-runtime-threats

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-protect-against-malware-zero-day-exploits-and-other-runtime-threats

How can I monitor the consumption of engines?

You can use the engine_metering_history information schema view for detailed tracking of FBU consumption.

Engines
COPY LINK TO ANSWER
how-can-i-monitor-the-consumption-of-engines-2

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-monitor-the-consumption-of-engines-2

How do I resolve 'Unable to cast' errors during CSV ingestion related to empty strings?

Use the NULLIF function to convert empty strings to NULL, which can then be cast to the appropriate data type.

When casting columns to data types like DATE or NUMERIC in Firebolt, empty strings in the source data can cause errors. This occurs because empty strings cannot be directly cast to other data types.

Use the NULLIF function to convert empty strings to NULL, which can then be cast to the appropriate data type without causing errors.

Example:

INSERT INTO tournaments_nullif_example_fact
SELECT
	NULLIF(dt, '')::date
FROM tournaments_nullif_example;

In this example, NULLIF(dt, '') converts empty strings in the dt column to NULL, allowing the data to be safely cast to a DATE type. This method ensures smooth casting of columns with empty strings in Firebolt.

ELT
COPY LINK TO ANSWER
how-do-i-resolve-unable-to-cast-errors-during-csv-ingestion-related-to-empty-strings

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-resolve-unable-to-cast-errors-during-csv-ingestion-related-to-empty-strings

What is Firebolt’s process for data retention and deletion?

Customers own their data and can delete it via commands like DROP DATABASE. Regardless, and upon contract termination, all customer data is deleted within 30 days.

Security
COPY LINK TO ANSWER
what-is-firebolts-process-for-data-retention-and-deletion

https://firebolt.io/faqs-v2-knowledge-center/what-is-firebolts-process-for-data-retention-and-deletion

I want to ensure that some of the engines in my account are accessible only to certain users. What mechanisms does Firebolt provide to help control what operations users can perform on engines?

Firebolt provides Role-based Access Control (RBAC) to help customers control which users can perform what operations on a given engine. For example, you can provide users with only the ability to use or operate existing engines but not allow them to create new engines. In addition, you can also prevent users from starting or stopping engines, allowing them to only run queries on engines that are already running. These fine-grained controls help ensure that customers do not end up with runaway costs resulting from multiple users in an organization creating and running new engines.

Engines
COPY LINK TO ANSWER
i-want-to-ensure-that-some-of-the-engines-in-my-account-are-accessible-only-to-certain-users-what-mechanisms-does-firebolt-provide-to-help-control-what-operations-users-can-perform-on-engines

https://firebolt.io/faqs-v2-knowledge-center/i-want-to-ensure-that-some-of-the-engines-in-my-account-are-accessible-only-to-certain-users-what-mechanisms-does-firebolt-provide-to-help-control-what-operations-users-can-perform-on-engines

How do I drop a table with indexes in Firebolt?

Drop the associated indexes manually or use the CASCADE option to automatically remove all dependencies when dropping the table.

ELT
COPY LINK TO ANSWER
how-do-i-drop-a-table-with-indexes-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/how-do-i-drop-a-table-with-indexes-in-firebolt

How does Firebolt handle external vulnerability reporting?

Researchers can report vulnerabilities by contacting security@firebolt.io.

Security
COPY LINK TO ANSWER
how-does-firebolt-handle-external-vulnerability-reporting

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-handle-external-vulnerability-reporting

How can I troubleshoot issues with Firebolt engines not starting due to low AWS instance availability?

If AWS instance availability is low:

Change the engine instance type.

Retry after some time.

Contact Firebolt support if the issue persists.

Engines
COPY LINK TO ANSWER
how-can-i-troubleshoot-issues-with-firebolt-engines-not-starting-due-to-low-aws-instance-availability

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-troubleshoot-issues-with-firebolt-engines-not-starting-due-to-low-aws-instance-availability

What backup and recovery options are available in case of Availability Zone (AZ) failure?

Customer data is stored in S3 buckets with high availability and durability. Our recovery objectives are:

- RTO (Recovery Time Objective): 12 hours

- RPO (Recovery Point Objective): 1 hour

- SLA (Service Level Agreement): 99.9%

Security
COPY LINK TO ANSWER
what-backup-and-recovery-options-are-available-in-case-of-availability-zone-az-failure

https://firebolt.io/faqs-v2-knowledge-center/what-backup-and-recovery-options-are-available-in-case-of-availability-zone-az-failure

How can I investigate query timeouts or unexpected delays in Firebolt engines?

To investigate query timeouts or delays, you can start by using Firebolt’s Query History and Query Profile tools, which provide detailed insights into query performance, including execution time, memory usage, and any potential bottlenecks. You can also check engine logs and metrics using Firebolt’s Engine Metrics History to identify issues like memory limitations, network latency, or resource constraints.

For troubleshooting steps, check the Firebolt documentation on query analysis.

Engines
COPY LINK TO ANSWER
how-can-i-investigate-query-timeouts-or-unexpected-delays-in-firebolt-engines

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-investigate-query-timeouts-or-unexpected-delays-in-firebolt-engines

Does Firebolt provide insurance coverage?

Yes, our insurance includes:

- Commercial General Liability

- Workers' Compensation and Employers' Liability

- Crime Insurance

- Professional & Technology Errors and Omissions

- Cyber Security Liability

Security
COPY LINK TO ANSWER
does-firebolt-provide-insurance-coverage

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-provide-insurance-coverage

What is spilling, and how does it work?

Spilling happens when a query requires more memory than allocated, causing intermediate query results to be stored on disk (SSD) instead of in-memory. While this ensures the query completes, it may affect performance.

For more information, check the Firebolt Documentation on Engine Metrics History.

Engines
COPY LINK TO ANSWER
what-is-spilling-and-how-does-it-work

https://firebolt.io/faqs-v2-knowledge-center/what-is-spilling-and-how-does-it-work

How to know the ACS URL for a certain organization?

Security
COPY LINK TO ANSWER
how-to-know-the-acs-url-for-a-certain-organization

https://firebolt.io/faqs-v2-knowledge-center/how-to-know-the-acs-url-for-a-certain-organization

What is the impact of stopping an engine on performance?

Stopping an engine in Firebolt results in the eviction of the local cache. This leads to a "cold start" upon restarting the engine, as queries initially must fetch data directly from storage, slowing down performance until the cache is replenished with frequently accessed data. To minimize performance degradation, consider pre-warming the engine with essential queries or data after it is restarted. For more information please check our documentation article Work with Engines using DDL.

Engines
COPY LINK TO ANSWER
what-is-the-impact-of-stopping-an-engine-on-performance

https://firebolt.io/faqs-v2-knowledge-center/what-is-the-impact-of-stopping-an-engine-on-performance

What is the system engine, and how is it used for metadata-related queries?

What is the system engine, and how is it used for metadata-related queries?The system engine in Firebolt is a lightweight, always-available engine specifically designed for metadata-related queries and administrative tasks. It supports various commands:

Access Control Commands: Manage roles, permissions, and users.

Metadata Commands: Execute queries on information schema views, such as information_schema.tables and information_schema.engines.

Non-Data Queries: Perform operations like SELECT CURRENT_TIMESTAMP() that do not involve table data.

Typical Use Cases:

Retrieve information about databases, tables, indexes, and engines.

Manage system configurations or user permissions.

Execute DDL operations like creating tables and views, and managing all engine-related operations (start, stop, drop, alter).

Engines
COPY LINK TO ANSWER
what-is-the-system-engine-and-how-is-it-used-for-metadata-related-queries

https://firebolt.io/faqs-v2-knowledge-center/what-is-the-system-engine-and-how-is-it-used-for-metadata-related-queries

Can engines be resized dynamically during operation?

Yes, Firebolt allows for dynamic resizing of engines during operation. You can adjust the number of nodes or the node type without stopping the engine, which lets workloads continue with minimal disruption. Use the ALTER ENGINE command to resize an engine. Newly started clusters post-resize will initially perform slower until they are warmed up.

Engines
COPY LINK TO ANSWER
can-engines-be-resized-dynamically-during-operation

https://firebolt.io/faqs-v2-knowledge-center/can-engines-be-resized-dynamically-during-operation

What happens to the queries that have already been run during an engine resize?

When an engine is resized dynamically, queries in execution will continue under the engine's original configurations until completion or until a timeout of 24 hours, after which they will be dropped if still running. The changes in engine size or type will only affect new queries submitted post-resize. Please check our documentation for more information.

Engines
COPY LINK TO ANSWER
what-happens-to-the-queries-that-have-already-been-run-during-an-engine-resize

https://firebolt.io/faqs-v2-knowledge-center/what-happens-to-the-queries-that-have-already-been-run-during-an-engine-resize

How can I check the last time an engine was started or stopped?

You can query the INFORMATION_SCHEMA.ENGINES table to check the last known status of an engine:

SELECT engine_name, last_started, last_stopped  
FROM INFORMATION_SCHEMA.ENGINES;

This will show when each engine was last started and stopped.

Engines
COPY LINK TO ANSWER
how-can-i-check-the-last-time-an-engine-was-started-or-stopped

https://firebolt.io/faqs-v2-knowledge-center/how-can-i-check-the-last-time-an-engine-was-started-or-stopped

Could you please give explanation, when it makes sense to scale engines with more clusters, when with higher amount of nodes and when with bigger engines?

Scaling with More Clusters:
This approach is ideal when you need to improve query concurrency—i.e., the ability to handle multiple queries simultaneously without significant performance degradation.

Scaling with a Higher Number of Nodes:
This is suitable when you find that the CPU utilization is consistently high, and queries are CPU-intensive. Adding more nodes spreads the workload across more computing units, thus alleviating CPU bottlenecks.

Scaling with Bigger Nodes:
This method is effective when the workload requires more memory or higher disk I/O capacity than what is currently available.

Engines
COPY LINK TO ANSWER
could-you-please-give-explanation-when-it-makes-sense-to-scale-engines-with-more-clusters-when-with-higher-amount-of-nodes-and-when-with-bigger-engines

https://firebolt.io/faqs-v2-knowledge-center/could-you-please-give-explanation-when-it-makes-sense-to-scale-engines-with-more-clusters-when-with-higher-amount-of-nodes-and-when-with-bigger-engines

Do we need to create new service account for new firebolt version or exist version will work fine?

If it's a new account, you'll need to set up a new user within that account, although this user can be linked to the existing service account.

If it's a new organization, you'll need to establish both a new service account and a new user within that organization and any associated accounts.

Deployment & Architecture
COPY LINK TO ANSWER
do-we-need-to-create-new-service-account-for-new-firebolt-version-or-exist-version-will-work-fine

https://firebolt.io/faqs-v2-knowledge-center/do-we-need-to-create-new-service-account-for-new-firebolt-version-or-exist-version-will-work-fine

Do you charge extra for AI features like vector search?

Firebolt’s AI-related features, such as vector search, are included within our standard pricing model. While these capabilities do utilize compute resources, there are no separate licensing fees or AI-specific upcharges. You only pay for the compute and storage you use, ensuring cost efficiency without hidden AI-related costs.

For a detailed breakdown of how AI workloads impact pricing, reach out to our team for a tailored estimate based on your use case.

AI
COPY LINK TO ANSWER
do-you-charge-extra-for-ai-features-like-vector-search

https://firebolt.io/faqs-v2-knowledge-center/do-you-charge-extra-for-ai-features-like-vector-search

How does Firebolt’s AI-powered data warehouse differ from competitors like Snowflake, Amazon Redshift, and Google BigQuery?

Firebolt is purpose-built for AI applications that require low-latency analytics. Unlike traditional cloud data warehouses, Firebolt delivers sub-second query performance for AI-driven workloads while supporting vector search and AI-driven optimizations. It enables faster and more efficient AI-powered analytics without the high costs and performance bottlenecks of legacy solutions.

AI
COPY LINK TO ANSWER
how-does-firebolts-ai-powered-data-warehouse-differ-from-competitors-like-snowflake-amazon-redshift-and-google-bigquery

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolts-ai-powered-data-warehouse-differ-from-competitors-like-snowflake-amazon-redshift-and-google-bigquery

Does Firebolt support real-time AI workloads?

Firebolt is optimized for low-latency, high-performance queries, but it is not a real-time processing platform. It excels in fast analytics on large-scale data but is not designed for event-driven streaming workloads. If your AI use case requires sub-second query execution, Firebolt is a great fit.

AI
COPY LINK TO ANSWER
does-firebolt-support-real-time-ai-workloads

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-support-real-time-ai-workloads

How does Firebolt’s vector search compare to dedicated vector databases?

Firebolt supports vector search but does not generate embeddings. Unlike dedicated vector databases, which specialize in unstructured data, Firebolt integrates vector search within a high-performance analytical data warehouse. This allows you to run hybrid queries (structured + unstructured) efficiently without managing separate systems.

If you already have embeddings generated from models like OpenAI, Hugging Face, or your own ML pipeline, Firebolt can store and query them at high speed and low latency, enabling AI-powered search and recommendations within your existing analytics environment.

AI
COPY LINK TO ANSWER
how-does-firebolts-vector-search-compare-to-dedicated-vector-databases

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolts-vector-search-compare-to-dedicated-vector-databases

How does Firebolt ensure low-latency performance for AI applications?

Firebolt is built on advanced indexing, vectorized query execution, and efficient storage optimizations, ensuring sub-second query performance even on large datasets. While it is not a real-time platform, Firebolt is ideal for AI-driven analytics, interactive dashboards, and personalized AI applications that demand ultra-fast queries on your data.

AI
COPY LINK TO ANSWER
how-does-firebolt-ensure-low-latency-performance-for-ai-applications

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-ensure-low-latency-performance-for-ai-applications

How can users ensure that their Firebolt service account setup is working correctly?

Users can verify the correct setup by ensuring they have obtained the API token, created the service account, and set up the appropriate automation steps. They can test ingestion by running queries to check if data has been successfully loaded into Firebolt.

Deployment & Architecture
COPY LINK TO ANSWER
how-can-users-ensure-that-their-firebolt-service-account-setup-is-working-correctly

https://firebolt.io/faqs-v2-knowledge-center/how-can-users-ensure-that-their-firebolt-service-account-setup-is-working-correctly

How can Firebolt users optimize query performance by leveraging primary indexes?

Firebolt users should define primary indexes based on frequently filtered columns, such as event dates and brand identifiers. By including relevant dimensions in the index, query performance can be significantly improved, as seen in the session where adding "brand" as a primary index reduced query execution time from minutes to seconds.

SQL
COPY LINK TO ANSWER
how-can-firebolt-users-optimize-query-performance-by-leveraging-primary-indexes

https://firebolt.io/faqs-v2-knowledge-center/how-can-firebolt-users-optimize-query-performance-by-leveraging-primary-indexes

How can users extract and save queries in Firebolt for collaboration?

Firebolt does not natively save queries across different user accounts. Users need to manually copy queries and store them externally, such as in Slack, Google Docs, or a shared repository, to ensure accessibility across their team.

SQL
COPY LINK TO ANSWER
how-can-users-extract-and-save-queries-in-firebolt-for-collaboration

https://firebolt.io/faqs-v2-knowledge-center/how-can-users-extract-and-save-queries-in-firebolt-for-collaboration

What are the best practices for structuring queries in Firebolt for performance optimization?

Users should avoid scanning large datasets unnecessarily by leveraging filtering on indexed columns. For example, filtering on both "event date" and "brand" significantly improves performance. Additionally, aggregating indexes can be used to precompute and store frequently used aggregations, reducing query execution time.

SQL
COPY LINK TO ANSWER
what-are-the-best-practices-for-structuring-queries-in-firebolt-for-performance-optimization

https://firebolt.io/faqs-v2-knowledge-center/what-are-the-best-practices-for-structuring-queries-in-firebolt-for-performance-optimization

How do aggregating indexes work in Firebolt, and what are their trade-offs?

Aggregating indexes in Firebolt store precomputed aggregations for faster query performance. They update automatically upon new data ingestion, reducing query execution time significantly. The trade-off is a slightly increased ingestion time since the indexes must be maintained.

SQL
COPY LINK TO ANSWER
how-do-aggregating-indexes-work-in-firebolt-and-what-are-their-trade-offs

https://firebolt.io/faqs-v2-knowledge-center/how-do-aggregating-indexes-work-in-firebolt-and-what-are-their-trade-offs

Does Firebolt provide an execution plan for queries, similar to Athena?

Yes, Firebolt provides a query execution plan. Users can view a visual representation after query execution or generate a text-based plan using the EXPLAIN command. This helps users compare performance against other engines like Athena.

SQL
COPY LINK TO ANSWER
does-firebolt-provide-an-execution-plan-for-queries-similar-to-athena

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-provide-an-execution-plan-for-queries-similar-to-athena

How can Firebolt be integrated with Apache Superset for visualization?

Firebolt provides documentation on connecting with Apache Superset. Their internal analytics team actively uses Superset, making it easier to provide support for any setup or troubleshooting questions.

Integrations
COPY LINK TO ANSWER
how-can-firebolt-be-integrated-with-apache-superset-for-visualization

https://firebolt.io/faqs-v2-knowledge-center/how-can-firebolt-be-integrated-with-apache-superset-for-visualization

How should large table joins be handled in Firebolt to optimize query performance?

In the discussion, the Firebolt team recommended creating new, pre-joined (or otherwise streamlined) tables rather than performing large, multi-table joins at query time. This approach, sometimes called "join elimination," can significantly reduce query overhead. In addition, the Firebolt team highlighted the importance of setting appropriate primary indexes on these new tables to further optimize performance.

SQL
COPY LINK TO ANSWER
how-should-large-table-joins-be-handled-in-firebolt-to-optimize-query-performance

https://firebolt.io/faqs-v2-knowledge-center/how-should-large-table-joins-be-handled-in-firebolt-to-optimize-query-performance

Does Firebolt always require creating additional tables, or can large joins be handled directly in SQL as with other warehouses (e.g., Snowflake)?

Firebolt does not strictly require more tables; however, to achieve high-performance queries on large datasets, many teams choose to create specialized tables with carefully designed primary indexes. Although Firebolt can perform joins directly in SQL, the discussion emphasized that pre-joining or restructuring certain data tables often yields better performance. This approach leverages Firebolt’s indexing and reduces the run-time cost of large joins.

SQL
COPY LINK TO ANSWER
does-firebolt-always-require-creating-additional-tables-or-can-large-joins-be-handled-directly-in-sql-as-with-other-warehouses-e-g-snowflake

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-always-require-creating-additional-tables-or-can-large-joins-be-handled-directly-in-sql-as-with-other-warehouses-e-g-snowflake

What is the recommended approach for migrating an existing Firebolt environment to a new organization or domain (for example, if a team is switching from one AWS org to another)?

Firebolt advises creating a brand-new organization under the desired domain or AWS account via the standard sign-up process. Once the new organization is set up, copy over any needed configurations, tables, or data from the old organization. Because a new organization starts as a new trial, you also receive fresh Firebolt usage credits. After verifying that everything works correctly in the new organization, the old one can be retired or deleted.

Deployment & Architecture
COPY LINK TO ANSWER
what-is-the-recommended-approach-for-migrating-an-existing-firebolt-environment-to-a-new-organization-or-domain-for-example-if-a-team-is-switching-from-one-aws-org-to-another

https://firebolt.io/faqs-v2-knowledge-center/what-is-the-recommended-approach-for-migrating-an-existing-firebolt-environment-to-a-new-organization-or-domain-for-example-if-a-team-is-switching-from-one-aws-org-to-another

Is it better to insert data into Firebolt one row at a time or in batches for real-time workloads?

Batch inserts are generally recommended for Firebolt. Inserting rows one at a time creates excessive overhead on the engine, leading to performance issues (especially on smaller engines). By sending records in small to moderate batches (e.g., once per second or at some reasonable time interval), the engine processes data more efficiently without overloading resources.

ELT
COPY LINK TO ANSWER
is-it-better-to-insert-data-into-firebolt-one-row-at-a-time-or-in-batches-for-real-time-workloads

https://firebolt.io/faqs-v2-knowledge-center/is-it-better-to-insert-data-into-firebolt-one-row-at-a-time-or-in-batches-for-real-time-workloads

How can organizations optimize queries that filter on high-cardinality timestamp columns (such as a ‘closed_at’ column)?

Use derived date columns (e.g., day-level granularity) to lower cardinality. For instance, store closed_at_day by truncating a TIMESTAMP to DATE. Incorporate the derived column (e.g., closed_at_day) into the primary index alongside other frequently used filters (e.g., tenant_id), allowing Firebolt to skip irrelevant data segments. Because raw timestamp columns can be extremely granular, indexing them directly often leads to poor selectivity. Restructuring the schema to include day- or hour-level columns can significantly improve performance. Leverage Firebolt’s caching features (result cache, sub-result cache) for repeated queries.

SQL
COPY LINK TO ANSWER
how-can-organizations-optimize-queries-that-filter-on-high-cardinality-timestamp-columns-such-as-a-closed-at-column

https://firebolt.io/faqs-v2-knowledge-center/how-can-organizations-optimize-queries-that-filter-on-high-cardinality-timestamp-columns-such-as-a-closed-at-column

Does Firebolt still require manual vacuuming, or is there an automatic process to reclaim storage space and optimize performance? Should vacuuming be done on a dedicated engine?

Firebolt has introduced an auto-vacuum feature that runs on the write engine. It triggers after a set number of transactions and reclaims space without blocking ingestion. Manual vacuuming is largely unnecessary in most use cases now. A dedicated engine is not typically required; auto-vacuum operates seamlessly on the existing write engine.

Engines
COPY LINK TO ANSWER
does-firebolt-still-require-manual-vacuuming-or-is-there-an-automatic-process-to-reclaim-storage-space-and-optimize-performance-should-vacuuming-be-done-on-a-dedicated-engine

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-still-require-manual-vacuuming-or-is-there-an-automatic-process-to-reclaim-storage-space-and-optimize-performance-should-vacuuming-be-done-on-a-dedicated-engine

How does Firebolt handle query performance as data within a single tenant expands?

Query performance primarily depends on the amount of data scanned. If queries remain selective (e.g., filtering by tenant ID and a truncated date range), Firebolt only scans relevant slices, keeping query times stable as data grows. Broad queries (e.g., SELECT * over wide date ranges) will naturally slow as more data must be scanned. Best practices include indexing on commonly used filters, leveraging caching, and avoiding unbounded queries to maintain good performance over time.

SQL
COPY LINK TO ANSWER
how-does-firebolt-handle-query-performance-as-data-within-a-single-tenant-expands

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-handle-query-performance-as-data-within-a-single-tenant-expands

How can organizations estimate the appropriate Firebolt engine size and associated costs for a given query load?

Firebolt offers multiple engine sizes (S, M, L, etc.). Smaller engines can handle many queries per minute if those queries are well-optimized. Heavier workloads or larger datasets may require a bigger engine or multiple concurrent engines. Firebolt charges by actual runtime (hourly or per second). Costs can be reduced by auto-stopping engines when not in use.

Pricing & Billing
COPY LINK TO ANSWER
how-can-organizations-estimate-the-appropriate-firebolt-engine-size-and-associated-costs-for-a-given-query-load

https://firebolt.io/faqs-v2-knowledge-center/how-can-organizations-estimate-the-appropriate-firebolt-engine-size-and-associated-costs-for-a-given-query-load

In the absence of true streaming, how can a Node.js application handle large Firebolt query results without running out of memory?

A viable workaround is to export query results to a file in S3, then read and process that file in smaller chunks. While it adds complexity (you must manage file paths, permissions, and cleanup), it avoids buffering the entire result set in application memory until real streaming is available in the Firebolt Node.js SDK.

Integrations
COPY LINK TO ANSWER
in-the-absence-of-true-streaming-how-can-a-node-js-application-handle-large-firebolt-query-results-without-running-out-of-memory

https://firebolt.io/faqs-v2-knowledge-center/in-the-absence-of-true-streaming-how-can-a-node-js-application-handle-large-firebolt-query-results-without-running-out-of-memory

What ingestion throughput can organizations expect, and how does Firebolt handle large batch loads or full refreshes?

Firebolt can ingest data at terabytes-per-hour scale, supported by internal benchmarks (e.g., half a terabyte in ~800 seconds on four S-sized engines). Actual throughput depends on factors such as file format, table schema, partitioning, and engine size. Organizations can scale up (larger engines or more engines) to accelerate big batch loads and scale down for smaller, more frequent delta loads.

ELT
COPY LINK TO ANSWER
what-ingestion-throughput-can-organizations-expect-and-how-does-firebolt-handle-large-batch-loads-or-full-refreshes

https://firebolt.io/faqs-v2-knowledge-center/what-ingestion-throughput-can-organizations-expect-and-how-does-firebolt-handle-large-batch-loads-or-full-refreshes

How should we structure the primary indexes?

Order columns by frequency of use in queries. For example, if tenant_id and closed_at_day appear in most filters, list them first. Within equally common columns, order from lowest cardinality to highest cardinality (fewest unique values to most). This approach ensures that Firebolt’s indexing effectively prunes unnecessary data scans for highly repetitive or frequently queried columns.

SQL
COPY LINK TO ANSWER
how-should-we-structure-the-primary-indexes

https://firebolt.io/faqs-v2-knowledge-center/how-should-we-structure-the-primary-indexes

Does Firebolt offer ongoing query-optimization assistance, and is there an extra cost associated with this service?

Firebolt’s Customer Success and Support teams provide complementary query optimization guidance—including index design, join performance tuning, and ingestion configuration—at no extra cost. Users can reach out via Slack or support tickets for best practices and troubleshooting. In many cases, optimization is a collaborative, ongoing process. If you notice a slower query, you can share it with Support; sometimes the solution is a schema or indexing change, other times a product fix may be required.

Support
COPY LINK TO ANSWER
does-firebolt-offer-ongoing-query-optimization-assistance-and-is-there-an-extra-cost-associated-with-this-service

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-offer-ongoing-query-optimization-assistance-and-is-there-an-extra-cost-associated-with-this-service

If we change the tenant ID, will the sub-result cache still be used?

In most cases, no. The sub-result cache benefits queries that overlap in the underlying data scanned or join results. If the tenant ID changes and there is little or no data overlap, the previous sub-results become irrelevant, so the cache will not offer a speed-up.

SQL
COPY LINK TO ANSWER
if-we-change-the-tenant-id-will-the-sub-result-cache-still-be-used

https://firebolt.io/faqs-v2-knowledge-center/if-we-change-the-tenant-id-will-the-sub-result-cache-still-be-used

Why do some queries run sub-second in Firebolt while others might take multiple seconds?

Queries with highly selective filters (e.g., smaller date ranges or high-selectivity columns) scan less data and often run in sub-second time. Queries that must scan large portions of the dataset (e.g., SELECT * over a broad date range) naturally take longer, especially on first (cold) runs when data must be read from storage. Firebolt’s sub-result caching reduces execution time for repeated or similar queries by caching portions of join results and aggregations. Proper indexing on commonly used filter columns can also significantly reduce the amount of data scanned, improving performance.

SQL
COPY LINK TO ANSWER
why-do-some-queries-run-sub-second-in-firebolt-while-others-might-take-multiple-seconds

https://firebolt.io/faqs-v2-knowledge-center/why-do-some-queries-run-sub-second-in-firebolt-while-others-might-take-multiple-seconds

What is the recommended approach for handling multiple workloads (read vs. write) in Firebolt? Should separate engines be used?

Firebolt is designed for a “decoupled compute” architecture where you can spin up separate engines for different workloads. A dedicated write engine handles ingestion, while one or more read engines handle queries. This ensures that write operations do not slow down queries and vice versa. You can also configure auto-start/auto-stop so that engines only run (and incur costs) when needed.

Deployment & Architecture
COPY LINK TO ANSWER
what-is-the-recommended-approach-for-handling-multiple-workloads-read-vs-write-in-firebolt-should-separate-engines-be-used

https://firebolt.io/faqs-v2-knowledge-center/what-is-the-recommended-approach-for-handling-multiple-workloads-read-vs-write-in-firebolt-should-separate-engines-be-used

Is continuous 24/7 ingestion engine usage required, or can the engine be started and stopped as needed?

Firebolt’s ingestion engine can be turned off when not actively loading data. Many teams schedule ingestion windows (e.g., hourly or daily) and then auto-stop the engine to save on costs. Billing is based on actual runtime, so you are not charged for idle ingestion clusters.

Deployment & Architecture
COPY LINK TO ANSWER
is-continuous-24-7-ingestion-engine-usage-required-or-can-the-engine-be-started-and-stopped-as-needed

https://firebolt.io/faqs-v2-knowledge-center/is-continuous-24-7-ingestion-engine-usage-required-or-can-the-engine-be-started-and-stopped-as-needed

Should we create separate Firebolt accounts for development, staging, and production, or use a single account with multiple databases?

Separate Accounts: Each account cleanly isolates its data and can map to separate AWS buckets or IAM roles. Single Account with Multiple Databases: Environments share the same account, so you must carefully permission each database. Most teams that maintain separate AWS resources (e.g., dev vs. staging vs. production buckets and roles) find it more straightforward to mirror that approach with separate Firebolt accounts.

Deployment & Architecture
COPY LINK TO ANSWER
should-we-create-separate-firebolt-accounts-for-development-staging-and-production-or-use-a-single-account-with-multiple-databases

https://firebolt.io/faqs-v2-knowledge-center/should-we-create-separate-firebolt-accounts-for-development-staging-and-production-or-use-a-single-account-with-multiple-databases

How often do we need to run vacuum if we do small, frequent updates—and does auto vacuum solve this?

Before auto vacuum, you would typically schedule vacuum after a certain number of inserts or on a time-based schedule (e.g., nightly). Firebolt’s auto vacuum feature (released around early 2025) automatically triggers a non-blocking vacuum every few hundred transactions in the background, substantially reducing or eliminating the need for manual vacuum scheduling. This occurs with minimal overhead and typically does not require an engine size increase.

ELT
COPY LINK TO ANSWER
how-often-do-we-need-to-run-vacuum-if-we-do-small-frequent-updates--and-does-auto-vacuum-solve-this

https://firebolt.io/faqs-v2-knowledge-center/how-often-do-we-need-to-run-vacuum-if-we-do-small-frequent-updates--and-does-auto-vacuum-solve-this

How does Firebolt support handle customer access, and can we restrict it?

Firebolt support engineers have the ability to access customer accounts for troubleshooting via Okta—only if they have the specific permissions. While it is generally recommended to keep support access open for fast incident resolution, you can request to block or limit their access if you have strict security requirements.

Security
COPY LINK TO ANSWER
how-does-firebolt-support-handle-customer-access-and-can-we-restrict-it

https://firebolt.io/faqs-v2-knowledge-center/how-does-firebolt-support-handle-customer-access-and-can-we-restrict-it

How can query performance be optimized when querying event data with minute-level granularity in Firebolt?

One approach is to restructure the table by setting the primary index on event_time to better leverage Firebolt’s indexing capabilities. Additionally, an aggregating index on event_time can be beneficial. However, if queries still take longer than expected (e.g., 15 seconds for 30 days of data), it may help to review: - The structure of the primary index and ensure it aligns with the query’s filtering. - Whether unnecessary dimensions are included in the dataset, increasing granularity unnecessarily. - If joins or aggregations can be optimized, possibly through pre-aggregated tables. Firebolt’s architecture is designed to improve query efficiency by avoiding costly full scans and optimizing indexing structures.

SQL
COPY LINK TO ANSWER
how-can-query-performance-be-optimized-when-querying-event-data-with-minute-level-granularity-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/how-can-query-performance-be-optimized-when-querying-event-data-with-minute-level-granularity-in-firebolt

Can Firebolt support a unified table for multiple reporting use cases (e.g., unique counts, injected data, and regular event data) instead of using multiple tables?

Yes, Firebolt can support a single table design that includes multiple reporting dimensions, such as unique counts, event times, and injected data. This consolidation can improve performance by reducing the need for complex joins and maintaining a single source of truth for analytics. However, when merging different data use cases into a single table, it is important to: - Optimize indexing to balance performance across different query patterns. - Consider partitioning or using aggregating indexes to precompute frequent aggregations. - Evaluate whether all reporting needs can be met within a single table without sacrificing efficiency.

SQL
COPY LINK TO ANSWER
can-firebolt-support-a-unified-table-for-multiple-reporting-use-cases-e-g-unique-counts-injected-data-and-regular-event-data-instead-of-using-multiple-tables

https://firebolt.io/faqs-v2-knowledge-center/can-firebolt-support-a-unified-table-for-multiple-reporting-use-cases-e-g-unique-counts-injected-data-and-regular-event-data-instead-of-using-multiple-tables

Is there a limit to how much data a single Firebolt engine can handle if I see a reference to a 1.8 TB size?

The “1.8 TB” figure refers to the SSD cache associated with a particular engine, not the total limit on data Firebolt can handle. Firebolt stores your full data in S3 for effectively unlimited capacity. Only the segments (tablets) relevant to a query are pulled into the SSD cache for faster processing. If your dataset exceeds 1.8 TB, Firebolt will still process it by cycling portions of data into and out of the SSD cache (first-in, first-out).

Engines
COPY LINK TO ANSWER
is-there-a-limit-to-how-much-data-a-single-firebolt-engine-can-handle-if-i-see-a-reference-to-a-1-8-tb-size

https://firebolt.io/faqs-v2-knowledge-center/is-there-a-limit-to-how-much-data-a-single-firebolt-engine-can-handle-if-i-see-a-reference-to-a-1-8-tb-size

How does connecting to the AWS Marketplace for billing work?

By subscribing through AWS Marketplace, you can consolidate Firebolt billing under your existing AWS billing arrangements. You will be directed to complete a few additional steps (“more clicks”) to finalize the purchase. Once completed, charges for your Firebolt usage appear in your AWS bill, simplifying vendor management if you prefer a single billing channel.

Pricing & Billing
COPY LINK TO ANSWER
how-does-connecting-to-the-aws-marketplace-for-billing-work

https://firebolt.io/faqs-v2-knowledge-center/how-does-connecting-to-the-aws-marketplace-for-billing-work

Does Firebolt have Generative AI features or an AI roadmap relevant to analytics use cases?

Firebolt has “GenAI” initiatives on its product roadmap. While exact capabilities may evolve, the published information highlights plans for: AI-Assisted Querying (e.g., query recommendations, natural language querying), Auto-Tuning & Optimization powered by machine learning, and Improved Developer Experience leveraging AI-based insights. For a deeper discussion of upcoming features, Firebolt can arrange a roadmap review session with its product team.

AI
COPY LINK TO ANSWER
does-firebolt-have-generative-ai-features-or-an-ai-roadmap-relevant-to-analytics-use-cases

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-have-generative-ai-features-or-an-ai-roadmap-relevant-to-analytics-use-cases

Is it possible to rename the organization URL (e.g., from shopware.firebolt.io to velo.firebolt.io)?

It is not clearly documented whether you can rename an existing organization URL. The typical workaround is to contact Firebolt support to see if they can rename it. If that is not feasible, you might need to recreate the organization under a new domain (e.g., using an email address at “velo”) and then migrate data or user setups.

Security
COPY LINK TO ANSWER
is-it-possible-to-rename-the-organization-url-e-g-from-shopware-firebolt-io-to-velo-firebolt-io

https://firebolt.io/faqs-v2-knowledge-center/is-it-possible-to-rename-the-organization-url-e-g-from-shopware-firebolt-io-to-velo-firebolt-io

If we use Change Data Capture (CDC) with very incremental updates, what concerns should we have about concurrency and vacuum tasks in Firebolt?

Concurrency & Overlapping Updates: If two CDC operations try to update the same row simultaneously, one transaction may fail. Implement a retry mechanism if you anticipate this scenario. Vacuum Operations: Frequent small inserts create multiple “tablets.” Vacuum consolidates and optimizes these for better query performance. Firebolt’s new “auto vacuum” (rolling out in early 2025) will greatly reduce the need for manual vacuum scheduling by automatically running a non-blocking vacuum in the background after a set number of transactions.

ELT
COPY LINK TO ANSWER
if-we-use-change-data-capture-cdc-with-very-incremental-updates-what-concerns-should-we-have-about-concurrency-and-vacuum-tasks-in-firebolt

https://firebolt.io/faqs-v2-knowledge-center/if-we-use-change-data-capture-cdc-with-very-incremental-updates-what-concerns-should-we-have-about-concurrency-and-vacuum-tasks-in-firebolt

How should we handle user management across different Firebolt accounts?

Organization Level (Authentication): You manage logins (email addresses) at the organization/workspace level. Account Level (Authorization): Each account defines its own users, roles, and permissions. A single login can exist in multiple accounts with different roles. Support Access: Firebolt support engineers can access accounts via Okta (with appropriate permissions), but you can opt to block this access if desired (not recommended).

Security
COPY LINK TO ANSWER
how-should-we-handle-user-management-across-different-firebolt-accounts

https://firebolt.io/faqs-v2-knowledge-center/how-should-we-handle-user-management-across-different-firebolt-accounts

Does Firebolt provide tools or capabilities to monitor database performance and scaling activities?

Yes, Firebolt provides monitoring capabilities through its information schema and metadata. Users are encouraged to implement custom monitoring and alerting processes on their side, although Firebolt also monitors performance and proactively alerts users to critical issues.

Monitoring & Performance
COPY LINK TO ANSWER
does-firebolt-provide-tools-or-capabilities-to-monitor-database-performance-and-scaling-activities

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-provide-tools-or-capabilities-to-monitor-database-performance-and-scaling-activities

Is there a way to Auto-Format or Beautify my query?

You can find "Format Script" when you click on the three dots on the SQL Tab.

SQL
COPY LINK TO ANSWER
is-there-a-way-to-auto-format-or-beautify-my-query

https://firebolt.io/faqs-v2-knowledge-center/is-there-a-way-to-auto-format-or-beautify-my-query

What are the considerations for splitting into separate Databases and Database best practices?

For users, it’s mainly about governance and logical isolation. Separate databases allow for different owners and permissions. Since custom schemas aren’t available yet, databases are the main way to group tables and views (this will change once schemas are supported).On the backend, metadata caching happens per database, so a single large database could add slight load. However, this is unlikely to have a practical impact unless in very large or complex cases.

Deployment & Architecture
COPY LINK TO ANSWER
what-are-the-considerations-for-splitting-into-separate-databases-and-database-best-practices

https://firebolt.io/faqs-v2-knowledge-center/what-are-the-considerations-for-splitting-into-separate-databases-and-database-best-practices

When should aggregating indexes be used in Firebolt, and what are their limitations?

Aggregating indexes in Firebolt pre-compute aggregated values to significantly speed up aggregation queries. They perform best when aggregations occur on a single fact table. They are less effective or infeasible when aggregation queries require multiple table joins because an aggregating index must be built on a single table only.

SQL
COPY LINK TO ANSWER
when-should-aggregating-indexes-be-used-in-firebolt-and-what-are-their-limitations

https://firebolt.io/faqs-v2-knowledge-center/when-should-aggregating-indexes-be-used-in-firebolt-and-what-are-their-limitations

Does Firebolt recommend separating ingestion and query engines, and why?

Yes. Firebolt recommends using separate engines for ingestion and query processing. Separating these concerns ensures ingestion tasks do not degrade query performance, leading to predictable and stable user experiences.

Engines
COPY LINK TO ANSWER
does-firebolt-recommend-separating-ingestion-and-query-engines-and-why

https://firebolt.io/faqs-v2-knowledge-center/does-firebolt-recommend-separating-ingestion-and-query-engines-and-why

What is the recommended approach for incremental data ingestion from PostgreSQL to Firebolt via AWS S3?

Firebolt recommends an incremental ingestion approach using S3 as a staging area. Data from PostgreSQL can be segmented (e.g., by ID range or time interval), pushed to S3, and loaded into Firebolt using the Firebolt SDK. This method ensures manageable load times and easy scaling by controlling the volume of data incrementally loaded.

ELT
COPY LINK TO ANSWER
what-is-the-recommended-approach-for-incremental-data-ingestion-from-postgresql-to-firebolt-via-aws-s3

https://firebolt.io/faqs-v2-knowledge-center/what-is-the-recommended-approach-for-incremental-data-ingestion-from-postgresql-to-firebolt-via-aws-s3

We use cookies to give you a better online experience
Got it