
Developers
Accelerate AI with Over 200 Supported Models, Effortlessly.
Discover how to quickly deploy your AI models on Rebellions' NPU using RBLN SDK.
You can find detailed information on our compiler, runtime, model zoo, and serving frameworks.
Get Started with Frameworks
RBLN SDK supports transformer and diffuser models on HuggingFace, downloadable from the Optimum RBLN library. Deploy newest models like Llama3-8b, SDXL from Huggingface Hub.
💡 Run HuggingFace models on Rebellions hardware.
- Compilation and inference with HuggingFace models optimized for Rebellions’ hardware.
- Efficient, developer-friendly API using RBLN Runtime.
- Support of Llama and SDXL models with multi chips.
RBLN SDK supports PyTorch 2.0. Accelerate your PyTorch-trained NLP, speech, and vision models on Rebellions’ hardware.
💡 RBLN SDK integrates PyTorch models.
- Compilation of PyTorch models optimized for Rebellions’ hardware.
- Efficient, developer-friendly API using RBLN Runtime.
- Run Torch 2.0 models without pretuning and build a powerful serving pipeline.
RBLN SDK supports TensorFlow. Optimize inference for models like LLMs, ImageNet and YOLO.
💡 RBLN SDK integrates TensorFlow models.
- Inference with a multitude of pre-trained Keras Applications.
- Efficient, developer-friendly API using RBLN Runtime.
- Run TensorFlow without pretuning and build a powerful serving pipeline.
Featured Resources
Rebellions specializes in the development of AI accelerators optimized to facilitate efficient AI inference across various advanced applications in fields such as finance and cloud computing.
Explore our latest documentation, tutorials, and webinars.
Rebellions’ Software Stack
Rebellions Software Stack supports our hardware to deliver maximum performance.
Machine Learning Framework
Compiler
Compute Library
Runtime Module
Driver
Firmware
RBLN Backend Rebellions Hardware
Rebellions’ Software Stack
Rebellions Software Stack supports our hardware to deliver maximum performance.
Rebellions Hardware
Frequently Asked Questions
To get started with RBLN SDK, download RBLN driver, compiler and Model Zoo in appropriate versions.
1. Install RBLN Driver.
2. Install RBLN Compiler.
3. Check whether RBLN SDK supports your desired model and check the code in Model Zoo.
4. Run the sample code to ensure there are no issues.
For a detailed installation guide, please refer to SDK Installation Guide.
We support the widely used Llama3, along with other Llama-based models (Solar, EEVE, etc.). Since the performance can be affected depending on the size and version of the models, please check Sample Code of each model.
Rebellions Model Zoo provides code examples for different models that you can optimize and test in our RBLN online documentation. You can find the code examples for each model on the following pages under the Task column:
- HuggingFace
- Pytorch
- Tensorflow
While we run tests covering most conceivable situations before release, bugs still happen.
Send us a file of Debug Dump Binaries (DDBs) so that we can provide technical support. The DDB file is safely encrypted and includes all the compilation steps and error logs. Please refer to Trouble and contact our Technical Support Team or email us.
While we aim to support most SOTA models on Hugging Face, we can’t guarantee that all of them will automatically run on Rebellions NPU.
Some models may require optimizations or improvements.
For any inquiries on models not listed on
Optimum RBLN, please contact our Technical Support Team or email us.
Yes. RBLN SDK provides a runtime interface bounded to C/C++, useful in environments where the Python Runtime is not supported or when applications need to achieve optimal execution time.
To use RBLN SDK in C/C++:
1. Update the APT repository
2. Install rbln sdk package
For more details, please refer to our documentation on C API.
The pipeline for model inference while running models on ATOM generally looks like this:
1. Preparing the pretrained model to run
2. Compiling the model with RBLN Compiler and saving the model
3. Loading the model with RBLN Runtime and running inference
4. Result (API etc)
Please note that the pipeline can differ depending on each service architecture. You can use RBLN SDK to run pre-trained deep learning models.
Yes, RBLN SDK supports Triton Inference Server and OpenAI API,
so that developers can focus on deploying models that meet the needs of their applications.
Yes. RBLN SDK supports configuration of multiple NPUs.
You can either use a single ATOM NPU to execute the model in parallel or divide the workload of a large model across multiple NPUs.
Currently, our Optimum RBLN supports this feature. For a list of supported models, please refer to Optimum RBLN.
Need help finding information?