LightAct 4 Docs
LightAct WebpagePDF Version
  • LightAct 4 Documentation
  • Understanding the Basics
    • Overview of Content Pipeline
    • Pushing Content out of LightAct
    • Terminology
    • How to map Content
      • Creating a Canvas
      • Creating a Layer
      • Creating a Video Screen
      • Mapping a Canvas to a Video screen
      • Creating a Perspective Thrower
      • Using Nodes to Set Content
      • Mapping to Outputs
    • Working quickly in LightAct
      • Window docking & workspaces
      • Drag & drop
      • Viewport Navigation
      • Shortcuts
    • Launching LightAct with Arguments
    • Top Bar
    • LightAct Performance
  • Transitioning to LightAct from other media servers
    • Switching to LightAct
    • LightAct for Disguise Operators
      • GUI Overview
      • Content Sequencing
      • Content Mapping
      • Projection Mapping
    • LightAct for Pixera Operators
  • Content mapping
    • Canvas mapping
    • Throwers
      • Thrower Properties
      • Perspective Thrower
      • Cubic thrower
      • Spherical Thrower
    • 3D model
      • Rendering Mode
      • Multiple materials & sources
      • Preparing 3D Models
        • Organize the 3D Scene
        • Materials in 3D models
        • UV Mapping of 3D Models
      • Outputting 3D Model's Texture
  • TIMELINES
    • Overview
    • Timelines Window
    • Timeline Editor Window
      • Sections
      • Markers
      • Curve Editor
    • Cues
      • Cue List
  • LAYERS AND LAYOUTS
    • Overview
    • Stock Layer Templates
      • Content Layer Templates
        • Video
      • Generative Layer Templates
        • Checkerboard
        • Color grid
        • Gradient
        • Mapping ID
        • Render Time
        • Scroll Texture
        • Send Texture
        • Solid Color
        • Strobe
        • Text
        • Texture to Mapping
      • Control Layer Templates
        • Go to Marker
        • Go to Section
        • Receive message
        • Send DMX
        • Send message
        • Set Fade
        • Set Timeline State
        • Set Volume
    • Layers
      • Layer Properties
      • Layer Layouts
      • Cross-fade
    • User Layer Templates
    • Variables
      • Variable Management
      • Control Variables with DMX, OSC, or Curves
    • Nodes
      • Node Connections
      • Node Action Flow
      • Order of Node Execution
      • Texture Processing
        • LUT
        • Luma Key
        • Rotate
      • Computer Vision
        • Optical Flow
        • Find Blobs
        • Mog2 Background Subtraction
        • Combine
        • Convert CV Color
        • Subtract
  • Node reference
    • Node reference
    • Layout nodes
      • Lifeline and actions node category
        • Tick node
      • Texture generators node category
    • Device nodes
  • Content Playback
    • Video Formats & Codecs
    • Playing Video
    • Playing Image Sequences
    • Images
    • Playing Notch Blocks
  • Projection Mapping
    • Projection Mapping
    • Projection Study
    • 2D Projection Mapping Workflow
    • 3D Projection Mapping Workflow
    • 3DCal
    • Iris
      • Setting up Iris Device Node
      • Adjusting Focus
      • Adding and calibrating Iris camera
      • Calibrating projectors using Iris
      • IrisVirtual
    • AutoBlend
    • Warp & Blend
      • Texture Warp & Blend node
      • Texture Mesh Warp node
      • Texture Perspective Warp node
      • Texture Softedge node
      • Warp & Blend Output Window
    • Texture Mask
  • DMX
    • DMX Overview
    • DMX Setup
    • DMX Out
      • Fixture Editor
      • Output to DMX Fixture
      • Set DMX in Layer Layouts
    • DMX In
      • Adjust layer variables with DMX
      • Index manager
      • Get DMX in Layer Layouts
      • Control by DMX
    • DMX by UV
  • Unreal Engine
    • Unreal Engine Integration Overview
    • Required Unreal Engine Plugins
      • Installing LightAct Plugin
      • Setting up the Plugin
        • LightAct Runtime Actor
        • LightAct Custom Time Step
        • LightAct Timecode
    • UnrealLink
    • Texture Sharing
      • Thrower2UCam
      • Projector2UCam
      • ViewportUCam
  • Integrations
    • Camera and object tracking
      • Camera Tracking
        • Stype
        • Mo-Sys
        • FreeD
        • Ncam
        • SPNet
      • Object Tracking
        • PSN
        • BlackTrax
        • OptiTrack
        • Vive
        • Antilatency
      • Tracking Visualizer
      • Tracking Follower
    • Network
      • TCP & UDP
      • OSC
        • OSCLearn
      • TUIO
    • Serial
      • MIDI
    • Audio
    • Content IO
      • NDI
      • Deltacast
        • FLEX Management
      • ZED
      • RealSense
    • OSCIn
    • 10-Bit Displays
  • LightTrack
  • LightNet
    • LightNet Overview
    • LightNet Clusters
      • Scan Network
      • Connect to Secondary Machines
      • Transfer and Open Project
      • Cluster Management
    • LightNet Performance
      • Canvas Distribution
      • Layer Distribution
      • Asset Distribution
      • Viewport Object Distribution
      • VRAM Improvements
    • LightSync
  • WebUI
    • WebUI Setup
  • GPU Management
    • EDID Management
      • Export EDID
      • Load EDID
      • Unload EDID
    • Synchronize GPU Outputs
    • Multiple GPUs
      • Mosaic
        • Set up Mosaic
        • Modify or Disable Mosaic
      • Primary Display
      • Workflow
    • DisplayPort - HDMI converters
  • Licensing
    • Licensing Overview
    • Adding or updating a License
    • Applying a License
    • Transferring a License
  • Troubleshooting
    • Iris Troubleshooting
    • Unreal Engine Integration Troubleshooting
      • Texture Sharing Troubleshooting
    • LightNet Troubleshooting
    • Notch Troubleshooting
    • GPU Management Troubleshooting
    • Asking for support
Powered by GitBook
On this page
  • Input pins
  • Output pins

Was this helpful?

  1. LAYERS AND LAYOUTS
  2. Nodes
  3. Computer Vision

Optical Flow

PreviousComputer VisionNextFind Blobs

Last updated 1 year ago

Was this helpful?

Optical Flow node analyzes the movement of pixels from frame to frame and detects moving parts based on the current and previous frames. This node requires CV Mat input and produces a CV Mat output.

Optical Flow node analyzes a CV image to identify moving pixels. It compares the current and previous input images to determine these moving parts. The resulting information is stored in the CV output as HSV values. The H channel represents the direction or displacement of the detected movement, while the V channel represents the velocity of the moving part.

Input pins

  • CV input: input image on which moving parts will be detected.

  • Scales number: the amount of detail in the moving area. A higher Scale numbers value translates into a more detailed Output image.

Good to know: If computation is slow, you can try lowering Scales number value.

  • Scales step: expressed as a ratio. If the object is small and moving fast through the images Scales number and Scales step parameters can be reduced.

  • Median filter size: higher number results in more "blurry" rectangular-shaped detected moving areas. If the object is moving fast through the image, the Median filter size parameter can be set to a higher number than the default.

  • Threshold magnitude: if the maximum detected magnitude, or velocity, in the whole image is below this value then the output image will be black.

  • Dynamic magnitude: If disabled then the magnitude for each pixel will be divided by Normalize magnitude parameter and stored in V channel. If this checkbox is Enabled, then the maximum magnitude will be set to 255 and the minimum magnitude will be set to 0 in V channel. All values in between will be linearly interpolated.

  • Dynamic Flow: If this checkbox is disabled, then only the current and previous images from the incoming video stream will be used in detection. If enabled, however, the optical flow algorithm will also use results from the previous detection.

Output pins

  • Output: detected parts of input images in HSV where H is moving direction and V is moving velocity.

  • Max magnitude: maximum magnitude, or velocity, of all detected parts in the Output image. This value is not normalized as is the V channel in the Output image.