A script to get Live Text from images
One of my favourite new features on Apple’s OSes in the last few years is Live Text, which is an optical character recognition tool that lets you select text in images. This sort of tech has been around for decades, but having it built into the OS makes it much easier to use. The text recognition isn’t perfect, but it works reliably and it’s good enough most of the time.
This is the promise of “it just works”: I open an image in Preview, hover my cursor over the text, I get a text selection cursor, then I right-click to get a menu of options:
This framework is also available for apps to use, and I found some instructions in an Apple article Recognizing Text in Images. I’ve been able to wrap this in a Swift script that lets me get live text on the command line. You can download the script from GitHub.
You pass the path to your image as a single command-line argument. Any text found in the image will be returned as a JSON list:
$ get_live_text railway-sign.jpg ["Passengers must","not pass this point","or cross the line"]
If the image doesn’t contain any text, it returns an empty list:
$ get_live_text dancers.jpg 
I’ve got a second script which scans a directory for images, gets Live Text for each of them, and saves the output to a JSON file.
I can think of a bunch of things I might do with this, starting with full-text search of images. (You can do search images by text if they’re in your Photos Library, but I have a lot of images which aren’t, including everything on my work laptop and my screenshot collection.)
Live Text may not be most accurate OCR tool, but for me it’s the most convenient. I’ve tried and failed to install command-line OCR in the past; getting this working took less than 15 minutes.