CyxWiz LogoCyxWiz
DocsNode Types

Node Types Reference

Complete reference for all 80+ node types available in the CyxWiz Node Editor.

Data Pipeline Nodes

Dataset InputData Pipeline

Load data from the Dataset Registry. First node in any training pipeline.

Outputs

  • data (Tensor) - Feature data
  • labels (Labels) - Target labels

Parameters

  • dataset_name - Name of loaded dataset
Data Loader

Batch iterator with shuffle and drop_last options.

ParameterDefaultDescription
batch_size32Samples per batch
shuffletrueShuffle each epoch
drop_lastfalseDrop incomplete batch
Data Split

Split data into train/validation/test sets.

ParameterDefaultDescription
train_ratio0.7Training set ratio
val_ratio0.15Validation set ratio
shuffletrueShuffle before split
stratifytrueMaintain class ratios

Core Layers

NodePurposeKey Parameters
DenseFully connected layerunits, use_bias, activation
Conv1D1D convolution for sequencesfilters, kernel_size, stride, padding
Conv2D2D convolution for imagesfilters, kernel_size, stride, padding, groups
Conv3D3D convolution for volumes/videoSame as Conv2D, but 3D
MaxPool2DMax poolingkernel_size, stride, padding
AvgPool2DAverage poolingkernel_size, stride, padding
GlobalMaxPoolGlobal max pooling-
GlobalAvgPoolGlobal average pooling-

Normalization Layers

BatchNorm

Batch normalization for training stability.

num_features, momentum=0.1, eps=1e-5, affine=true

LayerNorm

Layer normalization for transformers.

normalized_shape, eps=1e-5

GroupNorm

Group normalization.

num_groups=32, num_channels

InstanceNorm

Instance normalization.

Similar to BatchNorm

Recurrent Layers

NodeDescriptionParameters
RNNSimple RNN layerhidden_size, num_layers, bidirectional, dropout
LSTMLong Short-Term Memoryhidden_size, num_layers, bidirectional, proj_size
GRUGated Recurrent Unithidden_size, num_layers, bidirectional
EmbeddingToken embedding lookupnum_embeddings, embedding_dim, padding_idx

Attention & Transformer

MultiHeadAttention

Multi-head attention mechanism.

embed_dim=512, num_heads=8, dropout=0.0

SelfAttention

Self-attention (Q=K=V from single input).

TransformerEncoder

Full transformer encoder block.

d_model=512, nhead=8, dim_feedforward=2048, num_layers=6

PositionalEncoding

Add positional information.

d_model=512, max_len=5000, dropout=0.1

Activation Functions

NodeFormulaUse Case
ReLUmax(0, x)Most common, fast
LeakyReLUmax(0.01x, x)Prevent dying neurons
GELUx * Phi(x)Transformers
Swishx * sigmoid(x)EfficientNet
Sigmoid1/(1+e^-x)Binary output
Tanh(e^x-e^-x)/(e^x+e^-x)Range [-1,1]
Softmaxe^xi / sum(e^xj)Classification

Loss Functions

NodeUse CaseFormula
MSELossRegressionmean((y-y')^2)
CrossEntropyLossMulti-class-sum(y*log(y'))
BCELossBinary-[y*log(y') + ...]
L1LossRobust regressionmean(abs(y-y'))
HuberLossRobustMSE if small, L1 if large

Optimizers

NodeAlgorithmGood For
SGDStochastic Gradient DescentSimple, well-tuned
AdamAdaptive momentsMost tasks
AdamWAdam with weight decayTransformers
RMSpropRoot mean square propRNNs

Output Node

Required: Every training graph must end with an Output node.

Inputs

  • prediction (Tensor) - Model output
  • loss (Loss) - Loss value for training