Skip to main content

Pipeline Overview

Preview

Core operators (detection, classification, pose, segmentation, tracking) are stable. Cascade (op.foreach, op.croproi) and streaming APIs are still in development.

New here?

Start with the Quickstart for an overview, or Model Compilation if you need to compile a model first. This page is the full pipeline reference.

This guide explains how pipeline stages fit together. All examples use from axelera.runtime import op.

Getting Started

from axelera.runtime import op
import numpy as np

# Build a detection pipeline
pipeline = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),
)

# Run on an image (numpy array, HxWxC, uint8, BGR or RGB)
image = np.zeros((480, 640, 3), dtype=np.uint8) # replace with your image
detections = pipeline(image) # -> list[DetectedObject]

for det in detections:
print(f"{det.class_id.name}: {det.score:.0%} at {det.bbox}")

Pipeline Stages

Every inference pipeline follows the same pattern:

Image -> Transforms -> Model Inference -> Decode -> NMS -> Coordinate Transform -> Result Wrapper
(preprocess) (op.load) (parse (filter (to_image_space) (axdetection,
raw tensor overlaps) axpose, etc.)
output)
StageWhat it doesExample operator
TransformsPrepare image for model input (resize, normalize, etc.)op.letterbox(), op.totensor(), op.normalize()
Model InferenceRun the neural networkop.load('model.axm') or op.onnx_model('model.onnx')
DecodeParse raw tensor into structured arrayop.decode_detections(), op.decode_pose(), op.decode_segmentation()
NMSRemove duplicate overlapping detectionsop.nms()
Coordinate TransformMap from model input space to original image pixelsop.to_image_space()
Result Wrapper (optional)Convert array to typed Python objects. Without this step you get raw np.ndarray which is perfectly usable.op.axdetection(), op.axpose(), op.axsegmentation()

Pipeline Form vs Step-by-Step

There are two ways to use operators: pipeline form (op.seq) and step-by-step.

Pipeline form is the recommended approach. op.seq chains operators together into a pipeline. Frame context is managed automatically.

Calling pipeline.optimized() analyzes the chain and fuses adjacent operations (e.g., merging letterbox + totensor + normalize into a single step that runs inside the .axm model execution). This reduces memory copies and speeds up inference. The pipeline works without optimized() -- it just runs faster with it.

pipeline = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),
)
optimized = pipeline.optimized() # fuse ops for speed (optional)
detections = optimized(image)

Step-by-step gives maximum flexibility -- useful for debugging, mixing custom Python logic between operators, or inspecting intermediate values. You must manage the frame context manually with op.frame_context(image).

with op.frame_context(image):
x = op.letterbox(640, 640)(image)
x = op.totensor()(x)
x = op.load('yolov8n-coco.axm')(x)
print(f"Raw output shape: {x.shape}") # inspect intermediate
x = op.decode_detections(algo='yolov8', num_classes=80)(x)
x = op.nms()(x)
x = op.to_image_space()(x)
detections = op.axdetection(class_id_type=op.CocoClasses)(x)

Mixing both: You can use pipeline form for the model portion and step-by-step for custom surrounding logic:

model_pipeline = op.seq(
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
)

with op.frame_context(image):
preprocessed = op.letterbox(640, 640)(image)
preprocessed = op.totensor()(preprocessed)
detections_raw = model_pipeline(preprocessed)
# custom logic here ...
detections = op.to_image_space()(detections_raw)
result = op.axdetection(class_id_type=op.CocoClasses)(detections)

Detection Pipeline

pipeline = op.seq(
# Preprocessing
op.letterbox(640, 640), # Resize with padding, maintain aspect ratio
op.totensor(), # HWC uint8 -> CHW float32 [0,1]

# Model inference
op.load('yolov8n-coco.axm'), # Run model, returns raw tensor

# Postprocessing
op.decode_detections( # Parse raw tensor -> (N, 6) array
algo='yolov8', # [x0, y0, x1, y1, score, class_id]
num_classes=80,
confidence_threshold=0.25,
),
op.nms(iou_threshold=0.45, max_boxes=300), # Remove overlapping boxes
op.to_image_space(), # MODEL_PIXEL -> IMAGE_PIXEL coordinates
op.axdetection(class_id_type=op.CocoClasses), # -> list[DetectedObject]
)

# Use the pipeline
detections = pipeline(image) # list[DetectedObject]
for det in detections:
print(f"{det.class_id.name}: {det.score:.0%} at {det.bbox}")

Step-by-step equivalent (same result, but each step visible):

with op.frame_context(image):
x = op.letterbox(640, 640)(image)
x = op.totensor()(x)
x = op.load('yolov8n-coco.axm')(x)
x = op.decode_detections(algo='yolov8', num_classes=80, confidence_threshold=0.25)(x)
x = op.nms(iou_threshold=0.45, max_boxes=300)(x)
x = op.to_image_space()(x)
detections = op.axdetection(class_id_type=op.CocoClasses)(x)

Classification Pipeline

pipeline = op.seq(
op.resize(size=256, half_pixel_centers=True), # Resize smaller edge to 256
op.centercrop((224, 224)), # Center crop to model input
op.totensor(),
op.normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),

op.load('squeezenet1.0-imagenet.axm'),
op.axclassification(class_id_type=op.ImagenetClasses), # -> list[Classification]
op.topk(k=5), # -> top 5 classifications
)

Pose Pipeline

Pose detection adds keypoints to each detection. The keypoint columns flow through NMS naturally alongside the bounding box.

pipeline = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8npose-coco.axm'),
op.decode_pose(algo='yolov8', num_keypoints=17), # -> (N, 57) array
op.nms(iou_threshold=0.45, max_boxes=300),
op.to_image_space(keypoint_cols=range(6, 57, 3)), # Transform bbox AND keypoint coords
op.axpose(), # -> list[PoseObject]
)

The keypoint_cols parameter tells to_image_space which columns contain x-coordinates of keypoints (every 3rd column starting at 6), so they get mapped to image space too.


Segmentation Pipeline

Instance segmentation returns two outputs from the model: detections and prototype masks. This requires tuple data flow using par and itemgetter.

pipeline = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8nseg-coco.axm'),

# decode_segmentation returns a plain tuple: (detections, protos)
op.decode_segmentation(algo='yolov8', num_classes=80),

# Process detections and protos separately, then recombine
op.par(
op.seq(op.itemgetter(0), op.nms(iou_threshold=0.45)), # NMS on detections
op.itemgetter(1), # Pass protos through
),
# After unnamed par: detections and protos are separate positional args

op.par(
op.seq(op.pack(), op.itemgetter(0), op.to_image_space()), # Transform det coords
op.proto_to_mask(), # Compute masks from det+protos
),

op.axsegmentation(class_id_type=op.CocoClasses), # -> list[SegmentedObject]
)

Why par + itemgetter? decode_segmentation returns a tuple (detections, protos). We need to apply NMS only to detections while keeping protos intact. par runs two branches on the same input; itemgetter(0) extracts detections, itemgetter(1) extracts protos.

Step-by-step equivalent (often clearer for segmentation):

with op.frame_context(image):
x = op.letterbox(640, 640)(image)
x = op.totensor()(x)
det_raw, proto_raw = op.load('yolov8nseg-coco.axm')(x)
detections, protos = op.decode_segmentation(algo='yolov8', num_classes=80)(det_raw, proto_raw)
detections = op.nms(iou_threshold=0.45, max_boxes=300)(detections)
masks = op.proto_to_mask()(detections, protos)
detections = op.to_image_space()(detections)
segments = op.axsegmentation()(detections, masks)

Cascade Pipeline (foreach + croproi)

Work in progress: Cascade support (op.foreach, op.croproi) is not yet complete. The API shape shown below reflects the planned design but may change.

Cascade pipelines run a second-stage model on each detection from the first stage. foreach iterates over a list, and croproi extracts the image region for each detection.

pipeline = op.seq(
# First stage: detect objects
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),

# Second stage: classify each detected region
op.foreach(
'classifications', # Name for output field
op.croproi(property='bbox'), # Crop image region from detection bbox
op.resize(224, 224),
op.totensor(),
op.load('classifier.axm'),
op.axclassification(),
),
)
# Result: NamedTuple(input=[DetectedObject, ...], classifications=[Classification, ...])

Tracker Integration

Tracking adds persistent identity to detections across video frames. Place op.tracker() after op.axdetection():

pipeline = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),
op.tracker(algo='bytetrack'), # -> list[TrackedObject]
)

# TrackedObject has:
# .track_id - persistent ID across frames
# .state - new/tracked/lost/removed
# .tracked - the DetectedObject it matched
# .predicted_bbox - Kalman-filtered bbox

Supported algorithms: 'bytetrack' (default), 'oc-sort', 'sort', 'tracktrack'.


Model Formats

Two file formats are used for models:

.axm (Axelera Model) -- A compiled neural network for the Axelera AIPU. This is the output of deploy.py (the compilation step). When loaded with op.load(), it runs the model and returns raw numpy output tensors. You build the surrounding pipeline (preprocessing, decoding, NMS, etc.) yourself. See Model Compilation for how to produce .axm files from PyTorch, ONNX, or Ultralytics models.

.axe (Axelera Executable) -- A complete pipeline package. It is a ZIP archive containing a pipeline.toml (which describes the full operator chain) and an embedded .axm. When loaded with op.load(), you get a ready-to-run pipeline -- no need to add preprocessing or postprocessing.

op.load() handles both formats automatically:

# .axm -- you build the pipeline around it
detector = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'), # just the model, returns raw tensors
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),
)

# .axe -- everything is bundled, just call it
detector = op.load('yolov8n-coco.axe') # complete pipeline
detections = detector(image)

To save a pipeline you built as an .axe file for later reuse:

pipeline.save_axe('yolov8n-coco.axe')

Running a Pipeline

Input Formats

The primary input format is a numpy array (H, W, C), dtype uint8. Torch tensors and PIL Images are also accepted and converted automatically.

import numpy as np
import cv2

# From file
image = cv2.imread('image.jpg') # BGR, HxWx3, uint8

# From camera
ret, image = cap.read() # BGR by default with OpenCV

Color Handling

Models are typically trained on RGB images, but OpenCV reads BGR. There are two approaches:

Convert before the pipeline (explicit):

image_rgb = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
detections = pipeline(image_rgb)

Add color conversion inside the pipeline (self-contained):

pipeline = op.seq(
op.colorconvert('bgr', 'rgb'), # first operator converts color
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
...
)
detections = pipeline(image) # pass BGR directly

Output Types

Result wrappers (op.axdetection, op.axpose, etc.) are optional. They convert raw numpy arrays into typed Python objects (DetectedObject, PoseObject, etc.).

Without wrappers -- you get raw np.ndarray directly, which is perfectly fine for custom processing or when you want full control:

pipeline_raw = op.seq(
op.letterbox(640, 640),
op.totensor(),
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
op.nms(),
op.to_image_space(),
# no axdetection -- returns np.ndarray (N, 6)
)
raw = pipeline_raw(image) # np.ndarray: [x0, y0, x1, y1, score, class_id]

With wrappers -- you get typed objects with named attributes (det.bbox, det.score, det.class_id). Use wrappers when you want:

  • Cascade pipelines: op.foreach + op.croproi read the .bbox attribute from typed objects
  • Built-in rendering: typed objects have a .draw() method for visualization
  • Cleaner code: det.class_id.name instead of int(row[5])
detections = pipeline(image)   # list[DetectedObject]
for det in detections:
print(det.class_id, det.score, det.bbox)

For step-by-step use, frame_context tracks the original image for coordinate mapping:

with op.frame_context(image) as fc:
...
# fc.input is the original image
# fc.saved contains letterbox metadata

Supported Models

Currently supported model families:

  • YOLO object detection: YOLOv5, YOLOv7, YOLOv8, YOLOv9, YOLOv10, YOLO11, YOLO26
  • YOLO pose estimation: YOLOv8-pose, YOLO11-pose
  • YOLO instance segmentation: YOLOv8-seg, YOLO11-seg
  • Classifiers: Any model with softmax output (ImageNet-style)

More architectures (the model-zoo models) are coming soon.


Custom Operators

You can create custom operators by subclassing op.Operator and implementing __call__:

from axelera.runtime import op
import numpy as np

class ScaleScores(op.Operator):
"""Scale all detection scores by a constant factor."""
factor: float = 1.0

def __call__(self, detections: np.ndarray) -> np.ndarray:
result = detections.copy()
result[:, 4] *= self.factor # column 4 is score
return result

# Use in a pipeline
pipeline = op.seq(
op.load('yolov8n-coco.axm'),
op.decode_detections(algo='yolov8', num_classes=80),
ScaleScores(factor=0.9), # custom operator
op.nms(),
op.to_image_space(),
op.axdetection(class_id_type=op.CocoClasses),
)

Combinators Quick Reference

CombinatorSignatureWhat it doesWhen to use
op.seq(a, b, c)x -> a(x) -> b(...) -> c(...)Execute in order, pipe output to nextBuilding any pipeline
op.par(a, b)x -> (a(x), b(x))Run multiple ops on same inputProcessing tuple elements separately
op.foreach(name, ops...)[x1, x2] -> NamedTuple(input, name)Apply ops to each list elementCascade: second-stage model on each detection
op.pack()a, b -> (a, b)Collect positional args into tupleAfter unnamed par, before itemgetter
op.unpack()(a, b) -> a, bMark tuple for arg unpackingWhen operator returns tuple but next expects separate args
op.itemgetter(i)(a, b) -> a (if i=0)Extract element from tupleSelecting from decode output or par results
op.identity(x)x -> xPass through unchangedPlaceholder in par branches

Key rule: Unnamed par automatically unpacks its result. Named par (all operators have names) returns a NamedTuple that is NOT unpacked.