Chromascope is a director's finder for iPhone. It uses LiDAR depth and AI refinement to produce realistic shallow depth-of-field, paired with log color profiles that match real cinema cameras and lenses.

What's different

The lens, not just the focal length.

A finder should help you visualize the shot — more than just a phone camera with frame lines drawn on top. Chromascope simulates the optics and the sensor, so the image you frame actually matches the image you'll capture on shoot day.

Real depth of field.

LiDAR depth and neural refinement render shallow focus that matches your chosen lens + t-stop. Falloff tracks distance naturally, just like a real lens. Bokeh holds its shape on edges, hair, and bright points. Anamorphic modes even simulate optical characteristics and oval bokeh.

Real cinema color.

Log profiles modeled on ARRI, Sony, RED, Canon and more. What you capture in the finder matches the color space of a real camera. Use your own custom LUTS and they'll work exactly the same as they would on the actual thing, or export it flat to preview looks in your NLE or color software.

Also included

Everything you need. Nothing you don't.

In the app

On screen.

Chromascope gallery — single image view with metadata
Chromascope gallery — multi-select view

Beta

In private TestFlight.

Send us a note and we'll get you in on the next build.

Request TestFlight Access