LightAct 4 Docs
LightAct WebpagePDF Version
  • LightAct 4 Documentation
  • Understanding the Basics
    • Overview of Content Pipeline
    • Pushing Content out of LightAct
    • Terminology
    • How to map Content
      • Creating a Canvas
      • Creating a Layer
      • Creating a Video Screen
      • Mapping a Canvas to a Video screen
      • Creating a Perspective Thrower
      • Using Nodes to Set Content
      • Mapping to Outputs
    • Working quickly in LightAct
      • Window docking & workspaces
      • Drag & drop
      • Viewport Navigation
      • Shortcuts
    • Launching LightAct with Arguments
    • Top Bar
    • LightAct Performance
  • Transitioning to LightAct from other media servers
    • Switching to LightAct
    • LightAct for Disguise Operators
      • GUI Overview
      • Content Sequencing
      • Content Mapping
      • Projection Mapping
    • LightAct for Pixera Operators
  • Content mapping
    • Canvas mapping
    • Throwers
      • Thrower Properties
      • Perspective Thrower
      • Cubic thrower
      • Spherical Thrower
    • 3D model
      • Rendering Mode
      • Multiple materials & sources
      • Preparing 3D Models
        • Organize the 3D Scene
        • Materials in 3D models
        • UV Mapping of 3D Models
      • Outputting 3D Model's Texture
  • TIMELINES
    • Overview
    • Timelines Window
    • Timeline Editor Window
      • Sections
      • Markers
      • Curve Editor
    • Cues
      • Cue List
  • LAYERS AND LAYOUTS
    • Overview
    • Stock Layer Templates
      • Content Layer Templates
        • Video
      • Generative Layer Templates
        • Checkerboard
        • Color grid
        • Gradient
        • Mapping ID
        • Render Time
        • Scroll Texture
        • Send Texture
        • Solid Color
        • Strobe
        • Text
        • Texture to Mapping
      • Control Layer Templates
        • Go to Marker
        • Go to Section
        • Receive message
        • Send DMX
        • Send message
        • Set Fade
        • Set Timeline State
        • Set Volume
    • Layers
      • Layer Properties
      • Layer Layouts
      • Cross-fade
    • User Layer Templates
    • Variables
      • Variable Management
      • Control Variables with DMX, OSC, or Curves
    • Nodes
      • Node Connections
      • Node Action Flow
      • Order of Node Execution
      • Texture Processing
        • LUT
        • Luma Key
        • Rotate
      • Computer Vision
        • Optical Flow
        • Find Blobs
        • Mog2 Background Subtraction
        • Combine
        • Convert CV Color
        • Subtract
  • Node reference
    • Node reference
    • Layout nodes
      • Lifeline and actions node category
        • Tick node
      • Texture generators node category
    • Device nodes
  • Content Playback
    • Video Formats & Codecs
    • Playing Video
    • Playing Image Sequences
    • Images
    • Playing Notch Blocks
  • Projection Mapping
    • Projection Mapping
    • Projection Study
    • 2D Projection Mapping Workflow
    • 3D Projection Mapping Workflow
    • 3DCal
    • Iris
      • Setting up Iris Device Node
      • Adjusting Focus
      • Adding and calibrating Iris camera
      • Calibrating projectors using Iris
      • IrisVirtual
    • AutoBlend
    • Warp & Blend
      • Texture Warp & Blend node
      • Texture Mesh Warp node
      • Texture Perspective Warp node
      • Texture Softedge node
      • Warp & Blend Output Window
    • Texture Mask
  • DMX
    • DMX Overview
    • DMX Setup
    • DMX Out
      • Fixture Editor
      • Output to DMX Fixture
      • Set DMX in Layer Layouts
    • DMX In
      • Adjust layer variables with DMX
      • Index manager
      • Get DMX in Layer Layouts
      • Control by DMX
    • DMX by UV
  • Unreal Engine
    • Unreal Engine Integration Overview
    • Required Unreal Engine Plugins
      • Installing LightAct Plugin
      • Setting up the Plugin
        • LightAct Runtime Actor
        • LightAct Custom Time Step
        • LightAct Timecode
    • UnrealLink
    • Texture Sharing
      • Thrower2UCam
      • Projector2UCam
      • ViewportUCam
  • Integrations
    • Camera and object tracking
      • Camera Tracking
        • Stype
        • Mo-Sys
        • FreeD
        • Ncam
        • SPNet
      • Object Tracking
        • PSN
        • BlackTrax
        • OptiTrack
        • Vive
        • Antilatency
      • Tracking Visualizer
      • Tracking Follower
    • Network
      • TCP & UDP
      • OSC
        • OSCLearn
      • TUIO
    • Serial
      • MIDI
    • Audio
    • Content IO
      • NDI
      • Deltacast
        • FLEX Management
      • ZED
      • RealSense
    • OSCIn
    • 10-Bit Displays
  • LightTrack
  • LightNet
    • LightNet Overview
    • LightNet Clusters
      • Scan Network
      • Connect to Secondary Machines
      • Transfer and Open Project
      • Cluster Management
    • LightNet Performance
      • Canvas Distribution
      • Layer Distribution
      • Asset Distribution
      • Viewport Object Distribution
      • VRAM Improvements
    • LightSync
  • WebUI
    • WebUI Setup
  • GPU Management
    • EDID Management
      • Export EDID
      • Load EDID
      • Unload EDID
    • Synchronize GPU Outputs
    • Multiple GPUs
      • Mosaic
        • Set up Mosaic
        • Modify or Disable Mosaic
      • Primary Display
      • Workflow
    • DisplayPort - HDMI converters
  • Licensing
    • Licensing Overview
    • Adding or updating a License
    • Applying a License
    • Transferring a License
  • Troubleshooting
    • Iris Troubleshooting
    • Unreal Engine Integration Troubleshooting
      • Texture Sharing Troubleshooting
    • LightNet Troubleshooting
    • Notch Troubleshooting
    • GPU Management Troubleshooting
    • Asking for support
Powered by GitBook
On this page
  • Receiving NDI
  • Setup in Devices window
  • Setup in the Layouts window
  • Sending NDI
  • Setup in Devices window
  • Setup in the Layouts window
  • Sending and receiving video frames

Was this helpful?

  1. Integrations
  2. Content IO

NDI

Network Device Interface (NDI) is a high performance standard that allows anyone to use real time video on existing IP video networks.

PreviousContent IONextDeltacast

Last updated 11 months ago

Was this helpful?

In this chapter, we will go through the steps required to send or receive video via NDI.

Receiving NDI

Setup in Devices window

In order to create a device, right-click on the Devices window and choose NDI receiver.

NDI primarily sends/receives texture, but you can check additional settings in checkboxes, for example, check MetaData if you wish to send/receive metadata.

At the moment NDI audio sending/receiving is not supported.

Extra IPs parameter can be used to add more addresses for machines from which you wish to discover NDI sources. Parameter Group is used to filter who can see the specific stream. If, for example, the sender specifies a group name, then the receiver on the other end needs to specify the same group name in order to see the stream name by clicking on the Refresh list button. Choose the stream by clicking on a radio button and check Listen checkbox in order to start the stream.

Setup in the Layouts window

In order to use data from Ndi Receiver device node, we have to create Reference node in Layer Layouts.

If you are receiving a Texture through the Ndi Receiver device node, you can also see it on a Layer Layout by simply connecting the Reference node with a Get Texture node.

Sending NDI

Setup in Devices window

In order to create a device, right-click on the Layer Layouts window and choose Ndi Sender.

If you wish to send MetaData, you can check it under the Stream collapsible header. If Clock video is checked, then added video frames will be rate-limited to match the current framerate they are submitted at. Enter the Group name if you wish to send frames to the specific group of receivers.

At the moment NDI audio sending/receiving is not supported.

Setup in the Layouts window

In order to use data from Ndi Sender device node, we have to create Reference node in Layer Layouts.

It is also possible to change properties in Ndi Receiver device node from the Layer Layouts. For that case you can use Set Ndi Stream layer node. Just fill the stream parameters you wish to change during run-time and when boolean input Streaming receives True value, then parameters will be updated.

Sending and receiving video frames

We will now combine 2 setups from previous paragraphs and we will simulate sending and receiving video frames.

Let's set the texture for sending and at the same time we will render the received texture to the canvas.