Scailable AI Manager on AI-BLOX

To ensure that our AI manger provides the fastest and safest modular AI pipeline on any supported device, we don’t take supporting devices lightly; we want to ensure full support of the whole platform, and we need to be convinced that the selected hardware tailors to edge AI use-cases. So, today we are happy to announce that we added another OEM to our supported hardware list: AI-BLOX.

What’s AI-BLOX?

Since edge AI requires specific hardware, the rollout of edge AI applications is often challenging. Therefore AI-BLOX has developed modular hardware blox to accelerate your edge AI application rollout. AI-BLOX provides a unique, modular, hardware approach to providing Edge AI hardware. We have experimented extensively with their device(s) and they are sturdy, easy to setup, and enable extremely rapid iterations of edge AI solutions from a hardware perspective. The device also features colorful modules that are easily customizable to support any use case:

For additional information see:

Why does Scailable support AI-BLOX?

As you might have noticed, we are selective in the hardware we support. We need to ensure a highly efficient AI/ML pipeline, from I/O to model execution, on the full platform. Hence, the Scailable AI Manager must properly interface with all the hardware components on the device (accelerators etc.), and we need to be confident that the device is suited for specific edge AI solutions.

With their modular design, AI-BLOX ticks all of these boxes. Thus, after extensive testing, we decided to add them to our supported devices. The modular take on hardware that AI-BLOX brings to the table ideally aligns with our modular approach to edge AI software: AI-BLOX separates accelerators and physical I/O modularly from the surrounding platform making it easy to quickly iterate edge AI solutions. The Scailable AI Manager separates the AI inference pipeline both from the underlying hardware (no need to compile your models specifically for your target), and from the AI model itself. This allows for extremely fast iterations of AI models on the edge by providing a clear separation of the model development and (embedded) pipeline development cycles.

Where can I learn more?

If you are interested in running your Scailable supported edge AI application on AI-BLOX hardware, it is easy to get started:


More blogs

More news

Scailable supporting Seeed NVIDIA devices

We are excited to announce support for Seeed’s NVIDIA Jetson devices. The AI manager, and all our edge AI development tools, can now readily be used on Seeed devices. As we all know, edge AI solutions often start from creative ideas to optimize business processes with AI. These ideas evolve into Proof-of-Concept (PoC) phases, where […]

Read more

From your CPU based local PoC to large scale accelerated deployment on NVIDIA Jetson Orin.

Edge AI solutions often start with a spark of imagination: how can we use AI to improve our business processes? Next, the envisioned solution moves into the Proof-of-Concept (PoC) stage: a first rudimentary AI pipeline is created, which includes capturing sensor data, generating inferences, post-processing, and visualization of the results. Getting a PoC running is […]

Read more

Allxon and Scailable Unite to further Revolutionize Edge AI Deployment

In a groundbreaking development that promises to reshape the landscape of edge AI deployment, Allxon and Scailable have joined forces to unveil the Scailable AI Manager, a cutting-edge edge AI deployment middleware now available as a plugin in the Allxon Plugin Station. This strategic collaboration brings together the expertise of Allxon in image-based over-the-air (OTA) […]

Read more

Want to know more?

Visit our Resources Hub or contact us to learn how Scailable can improve your business operations.

Contact us

Interested in our platform?

Get a free demo of our AI Manager so you can experience the benefits of edge AI solutions.

Book a demo