LightAct 4 Docs
LightAct WebpagePDF Version
  • LightAct 4 Documentation
  • Understanding the Basics
    • Overview of Content Pipeline
    • Pushing Content out of LightAct
    • Terminology
    • How to map Content
      • Creating a Canvas
      • Creating a Layer
      • Creating a Video Screen
      • Mapping a Canvas to a Video screen
      • Creating a Perspective Thrower
      • Using Nodes to Set Content
      • Mapping to Outputs
    • Working quickly in LightAct
      • Window docking & workspaces
      • Drag & drop
      • Viewport Navigation
      • Shortcuts
    • Launching LightAct with Arguments
    • Top Bar
    • LightAct Performance
  • Transitioning to LightAct from other media servers
    • Switching to LightAct
    • LightAct for Disguise Operators
      • GUI Overview
      • Content Sequencing
      • Content Mapping
      • Projection Mapping
    • LightAct for Pixera Operators
  • Content mapping
    • Canvas mapping
    • Throwers
      • Thrower Properties
      • Perspective Thrower
      • Cubic thrower
      • Spherical Thrower
    • 3D model
      • Rendering Mode
      • Multiple materials & sources
      • Preparing 3D Models
        • Organize the 3D Scene
        • Materials in 3D models
        • UV Mapping of 3D Models
      • Outputting 3D Model's Texture
  • TIMELINES
    • Overview
    • Timelines Window
    • Timeline Editor Window
      • Sections
      • Markers
      • Curve Editor
    • Cues
      • Cue List
  • LAYERS AND LAYOUTS
    • Overview
    • Stock Layer Templates
      • Content Layer Templates
        • Video
      • Generative Layer Templates
        • Checkerboard
        • Color grid
        • Gradient
        • Mapping ID
        • Render Time
        • Scroll Texture
        • Send Texture
        • Solid Color
        • Strobe
        • Text
        • Texture to Mapping
      • Control Layer Templates
        • Go to Marker
        • Go to Section
        • Receive message
        • Send DMX
        • Send message
        • Set Fade
        • Set Timeline State
        • Set Volume
    • Layers
      • Layer Properties
      • Layer Layouts
      • Cross-fade
    • User Layer Templates
    • Variables
      • Variable Management
      • Control Variables with DMX, OSC, or Curves
    • Nodes
      • Node Connections
      • Node Action Flow
      • Order of Node Execution
      • Texture Processing
        • LUT
        • Luma Key
        • Rotate
      • Computer Vision
        • Optical Flow
        • Find Blobs
        • Mog2 Background Subtraction
        • Combine
        • Convert CV Color
        • Subtract
  • Node reference
    • Node reference
    • Layout nodes
      • Lifeline and actions node category
        • Tick node
      • Texture generators node category
    • Device nodes
  • Content Playback
    • Video Formats & Codecs
    • Playing Video
    • Playing Image Sequences
    • Images
    • Playing Notch Blocks
  • Projection Mapping
    • Projection Mapping
    • Projection Study
    • 2D Projection Mapping Workflow
    • 3D Projection Mapping Workflow
    • 3DCal
    • Iris
      • Setting up Iris Device Node
      • Adjusting Focus
      • Adding and calibrating Iris camera
      • Calibrating projectors using Iris
      • IrisVirtual
    • AutoBlend
    • Warp & Blend
      • Texture Warp & Blend node
      • Texture Mesh Warp node
      • Texture Perspective Warp node
      • Texture Softedge node
      • Warp & Blend Output Window
    • Texture Mask
  • DMX
    • DMX Overview
    • DMX Setup
    • DMX Out
      • Fixture Editor
      • Output to DMX Fixture
      • Set DMX in Layer Layouts
    • DMX In
      • Adjust layer variables with DMX
      • Index manager
      • Get DMX in Layer Layouts
      • Control by DMX
    • DMX by UV
  • Unreal Engine
    • Unreal Engine Integration Overview
    • Required Unreal Engine Plugins
      • Installing LightAct Plugin
      • Setting up the Plugin
        • LightAct Runtime Actor
        • LightAct Custom Time Step
        • LightAct Timecode
    • UnrealLink
    • Texture Sharing
      • Thrower2UCam
      • Projector2UCam
      • ViewportUCam
  • Integrations
    • Camera and object tracking
      • Camera Tracking
        • Stype
        • Mo-Sys
        • FreeD
        • Ncam
        • SPNet
      • Object Tracking
        • PSN
        • BlackTrax
        • OptiTrack
        • Vive
        • Antilatency
      • Tracking Visualizer
      • Tracking Follower
    • Network
      • TCP & UDP
      • OSC
        • OSCLearn
      • TUIO
    • Serial
      • MIDI
    • Audio
    • Content IO
      • NDI
      • Deltacast
        • FLEX Management
      • ZED
      • RealSense
    • OSCIn
    • 10-Bit Displays
  • LightTrack
  • LightNet
    • LightNet Overview
    • LightNet Clusters
      • Scan Network
      • Connect to Secondary Machines
      • Transfer and Open Project
      • Cluster Management
    • LightNet Performance
      • Canvas Distribution
      • Layer Distribution
      • Asset Distribution
      • Viewport Object Distribution
      • VRAM Improvements
    • LightSync
  • WebUI
    • WebUI Setup
  • GPU Management
    • EDID Management
      • Export EDID
      • Load EDID
      • Unload EDID
    • Synchronize GPU Outputs
    • Multiple GPUs
      • Mosaic
        • Set up Mosaic
        • Modify or Disable Mosaic
      • Primary Display
      • Workflow
    • DisplayPort - HDMI converters
  • Licensing
    • Licensing Overview
    • Adding or updating a License
    • Applying a License
    • Transferring a License
  • Troubleshooting
    • Iris Troubleshooting
    • Unreal Engine Integration Troubleshooting
      • Texture Sharing Troubleshooting
    • LightNet Troubleshooting
    • Notch Troubleshooting
    • GPU Management Troubleshooting
    • Asking for support
Powered by GitBook
On this page
  • Adding a camera to the Viewport
  • Mapping Iris camera
  • Mapping Iris camera to Iris device node
  • Mapping projection objects to Iris camera
  • Calibrating a camera

Was this helpful?

  1. Projection Mapping
  2. Iris

Adding and calibrating Iris camera

Adding an Iris camera to the viewport and calibrating it

PreviousAdjusting FocusNextCalibrating projectors using Iris

Last updated 3 years ago

Was this helpful?

Adding a camera to the Viewport

Insert an Iris camera and place it roughly where its physical counterpart is relative to the projection object. You also might want to adjust its frustum, so that it corresponds to the FOV of the physical camera.

The easiest way to is to use Lens override dropdown.

Mapping Iris camera

Before we can proceed to calibrating Iris camera, we need to map it to Iris device node and to the projection objects.

Mapping Iris camera to Iris device node

Select Iris camera in Viewport. In its properties on the right, in Device section, select Iris device node you created in previous chapter.

Good to know: Mapping Iris camera in the Viewport to an Iris device node tells Iris camera which physical Iris camera it should get its image from.

Mapping projection objects to Iris camera

Mapping of projection objects to Iris camera in the Viewport is done either by dragging from projection object to the camera or by using the camera's Sources properties.

Once this is done, the camera should see theprojection object in the Viewport and the image of the virtual object should appear when you select the camera.

Calibrating a camera

To calibrate Iris camera, just double click on the camera in the viewport or right click on it and select Setup.

Iris’ 3DCal window will open and the process of calibrating the camera is more or less the same as the process of calibrating a projector. Therefore, we won’t go into the details here, but please refer to 3DCal chapter:

The only difference between the 2 windows are these.

  • In the 3DCal window of the projector, you select Projected points on the right hand side. In the camera’s window, you select Physical points and you click on the camera’s image itself (as opposed to clicking on the cursor on the projected object).

  • Camera’s 3DCal window has Stream section in the properties, where you can adjust the settings of the device node.

  • It also has Overlay opacity which is very helpful in order to visually judge the precision of the calibration process.

After you’ve selected 6 or more points camera will get calibrated. Below you will see an example of a good calibration of the camera.

You can see, the reprojection error is below 1px and the yellow calibration lines are all parallel.

Another useful indicator of a good calibration, is changing the Overlay opacity (you can use shortcut ‘O’). If, when changing the Overlay opacity, the visual and physical image don’t change much, it means that the virtual camera is precisely where the physical camera is in relation to the projection object.

3DCal