Getting Started
The onboarding wizard walks you through creating your organization and workspace in four steps. Everything is collected up front and submitted together at the end.
Sign in at optimaflo.io/sign-in with Google, Microsoft, Amazon, or email
Organization: name your organization, set a URL slug, and select your industry and team size
Workspace: create one or more workspaces (up to 5 during onboarding). Separate by environment, team, or project.
BYOC: configure your Bring Your Own Cloud deployment on GCP. Select your infrastructure tier and region, then the platform provisions everything. AWS and Azure support is coming soon.
Each workspace gets its own Apache Iceberg catalog, so data from different workspaces is fully isolated — even within the same organization.
OptimaFlo currently supports Google Cloud Storage, BigQuery, and REST API connectors, with more on the way. The Data Source AI agent guides you through the entire process conversationally; tell it what you want to connect and it handles authentication, browsing, file selection, validation, and schema inference.
Open Data Sources in the sidebar and click Add Source — this opens the Data Source AI agent
Tell the agent what you want to connect (e.g. "Connect my GCS bucket gs://company-data") — it authenticates via OAuth for cloud sources or asks for credentials for databases
The agent browses your buckets, folders, or tables and lets you select specific files or datasets to ingest
It validates the connection, infers your schema, and creates the data source record — ready for use in a pipeline
For GCS and BigQuery, authentication happens via a Google OAuth popup — no service account keys to manage. The platform handles token refresh automatically.
Three ways to build: drag-and-drop on the visual canvas, use the Pipeline Generator AI to create a pipeline from a prompt, or let Concierge AI handle the entire workflow end-to-end.
"Connect my GCS bucket gs://sales-data, clean the CSVs, deduplicate on order_id, and create a monthly revenue summary."
Execute your pipeline manually or set a schedule. OptimaFlo generates an Apache Airflow DAG behind the scenes; you never touch Airflow config directly.
Click Execute to run the pipeline immediately. The platform auto-saves before executing
Monitor progress in the execution panel and watch each layer complete from Ingestion through Aggregation
Open Settings to set a schedule (hourly, daily, weekly, monthly, quarterly, or yearly). The platform converts your selection to an Airflow cron
Use Backfills to re-process historical date ranges when you change transformation logic
Backfills run sequentially by default to avoid Iceberg write conflicts. You can increase parallelism for independent tables.
Your source data lands here untouched. Every record is preserved in Apache Iceberg tables with full history, ACID transactions, and time-travel.
Cleaned, deduplicated, and type-cast. SQL Copilot generates transformations from plain English, then you review and approve before anything runs.
Aggregated, business-ready metrics and star schemas. Feed dashboards, exports, and analyst queries from a single source of truth.
Connect a source, transform with SQL, and schedule with Airflow — all from the browser.
AI-native data platform. From raw data to business dashboards powered by Apache open standards, visual pipeline building, and AI agents that handle the heavy lifting.
© 2026 OptimaFlo. All rights reserved.
We use cookies to enhance your browsing experience, serve personalized content, and analyze our traffic. By clicking "Accept All", you consent to our use of cookies. You can customize your preferences or learn more in our Cookie Policy and Privacy Policy.