IOblend

technology

making it work

office desk, man, business-6952919.jpg

Build and run production-grade Apache Spark data pipelines

Powerful, versatile and simple - one tool for all data integration jobs

We believe in simplicity and versatility of data integration tools. This is why we created a “Swiss army knife” solution to allow you to do all data integration jobs with just a single tool.

You will drastically reduce the effort and cost of production-grade ETL development, multiple tool ecosystem maintenance and manual data wrangling. 

If you want to get the most out of your valuable data or deploy the power of AI fast, choose IOblend.

Technical highlights

Uses Apache Spark framework

Real-time, production grade, managed Apache Spark™ data pipelines in minutes. Easy-to-use designer and a powerful engine for any data pipeline architecture (ETL/ELT/ReETL).

Automatic in-built DataOps

Record-level lineage, CDC, metadata, schema, eventing, de-duping, SCD, chained aggregations, MDM, cataloguing, regressions, windowing, compaction – all features are in-built and require no additional Spark coding.

Easy integration of real-time streaming (transactional event) and batch data

IOblend is build around Kappa architecture, which enables easy handling to both batch and real-time data, saving you time and cost. 

Low code / no code development

Low code development without the usual downsides.

We specifically made sure you could use IOblend for any data integration job, no matter the complexity. We just abstracted away the coding complexity associated with Spark to reduce the development burden.

Use SQL or Python for data transformations to handle your business rules, specific quality policies and other constraints.

IOblend will automatically build, run and manage efficient Spark pipelines in the background for you.

puzzle pieces, puzzle, share-7696621.jpg

Applicable for all data integration use cases

No data challenge is beyond reach with IOblend

Data migration, system integration, real-time IoT analytics, central or federated data architectures, data synchronisation, from simple ingests to full end-to-end data processing in-flight – IOblend caters to all data integration use cases.

Work with any data wherever it resides

Real-time integration with Snowflake, GCP, AWS, Salesforce, Oracle, Databricks, Microsoft Azure, SAP plus many more.

IOblend connects to all data sources and sinks via JDBC, ODBC, API, ESB, dataframes or flat files.

Low latency, massively parallelised data processing

IOblend was designed with real-time data applications in mind.

We have optimised Spark for extreme performance on modest machinery to improve computational efficiencies and reduce cost (we have tested at >10m transactions per sec).

Flexible and cost-effective deployment

IOblend runs on any environment – local, on-prem, edge, cloud and hybrid. 

Storage and compute are decoupled, so you can deploy the processing engine independently of your data repositories for best performance and cost.

Example: source the data from your on-prem Oracle ERP, process it using AWS EC2 or EMR, push the results into MS Azure for analytics. 

Watch IOblend in action

See how easy it is to build and run Apache Spark data pipelines with IOblend.

Product Demo

data_engineering_spark

How it works

IOblend has two core components

IOblend Designer and IOblend Engine.

IOblend Designer is a desktop GUI for interactively designing, building and testing data pipeline DAGs. This process produces IOblend metadata describing the data pipelines that need to be executed.

IOblend Engine is the heart of IOblend that takes data pipeline metadata and converts it into Spark streaming jobs to be executed on any Spark cluster.

Metadata driven development

IOblend data pipelines are defined in playbooks, which store configuration, run parameters and the business logic. The playbook components are stored as JSON files and can be easily reused and shared among developers to speed up further development work.

The IOblend Engine dynamically converts playbook information to Apache Spark™ streaming jobs and executes them efficiently without the need to code.

IOblend Engine is installed on any Cloud or on-prem environment as either a Linux container or directly into your Spark infrastructure such as Databricks, HDIsight, Google Cloud Proc or AWS EMR as well as on-prem Spark infrastructure.

IOblend resides entirely inside the client’s environment, inheriting security protocols for a complete peace of mind.

Do development and production running with a single tool

IOblend can be used to develop and run data pipelines

The intuitive Designer front end interface allows for easy, interactive data pipeline development.

You can add/remove as many dataflow components as required, build and test them one at a time, in sequence or its entirety, and create fully productionised data pipelines.

We made it very easy to build real-time streaming and batch data pipelines

Create advanced dataflows for streaming or batch data from any source and sink to any environment prem/cloud/hybrid environments – and push your data back to the source just as easily if required.

We have templated and annotated the Designer options to help you develop data pipelines faster.

Development mode

In Dev mode, we have added a visual interface to let you test and inspect each step of your data pipeline development as you progress.

Pause, amend and update your sources, transforms and sinks while running dataflows live, without the need to stop the job while you work on it.

You can easily see and export the schemas.

Data Pipelines IOblend modular design
Ask us all about IOblend

Request an in-depth technical discussion and demo today.

call center, headset, woman-8643476.jpg

IOblend - the basics

IOblend has two core components: IOblend Designer and IOblend Engine

IOblend Designer is a desktop GUI for interactively designing, building and testing data pipeline DAGs. This will produce the IOblend metadata describing the data pipelines that need to be executed

IOblend Engine: The heart of IOblend that takes IOblend data pipeline metadata and converts it into spark streaming jobs to be executed on any Spark cluster

IOblend comes in two flavours: IOblend Developer Suite and IOblend Enterprise Suite

IOblend Developer Suite:

Both the IOblend Designer and IOblend Engine are installed on a developer’s desktop/laptop. IOblend automatically creates a local Spark environment for the IOblend Engine to work with and executes pipelines created by IOblend Designer

Although the IOblend engine is running on your local machine, it can still connect to any Cloud or On-prem source you have access to pull the data from and use within your data pipeline. Likewise, it also writes the results of a data pipeline to any Cloud or On-prem structure that you have access to. Data pipelines can only be executed from the Designer

IOblend Enterprise Suite:

Similar to the Development Suite but this time you have a remote IOblend Engine, packaged to run on any Cloud or On-prem Spark environment. IOblend Designer will generate ‘run’ files that can be executed/scheduled to run in your enterprise Spark infrastructure. IOblend designer can be connected to either the local IOblend Engine or the remote IOblend engine for development and testing. The IOblend run files can be scheduled by any scheduling software such as Airflow.

IOblend supports collaborative development and pipeline versioning.

All IOblend data pipelines are stored as JSON metadata files, which means they can be placed in any code repository and versioned, just like standard software development.

IOblend deployment architecture

Local deployment

Local machine deployment is best suited for dataflow development

IOblend ships with a containerised Spark environment, so it works out of the box – no need for developers, analysts and data scientists to build local Spark environments

The software connects to the client’s data systems via the existing security protocols, so no data is ever exposed externally

ioblend_dataops_deployement_local

Cloud deployment

ioblend_dataops_cloud_deployment

Only the Designer is installed on the local machine

IOblend Engine is installed in your Cloud or on-prem environment as either a Linux container or directly into your Spark infrastructure such as Databricks, HDIsight, Google Cloud Proc or AWS EMR, as well as on-prem Spark infrastructure

The Linux image contains all the essential IOblend dataflow components within

The user can easily interact and run dataflows from within the Cloud to any external systems (on prem)

IOblend supports all major Cloud systems and allows you to work with multiple Clouds simultaneously

Scroll to Top

Attention Data Developers

FREE DEVELOPER EDITION

Download your FREE copy and see how easy it is to start building production grade Spark data pipelines with IOblend