I keep thinking about Tom Ohanian’s series on the State of Digital Nonlinear Editing. Specifically these paragraphs in Part 10:
Content that is recorded will then be processed by a variety of AI application suites. Each suite will provide different functionality (e.g. tonal analysis, speech-to-text, etc.) based on the characteristics of the content. … Very rich, detailed, and comprehensive metadata about that content will result without the large number of humans currently associated with these tasks.
At that point, the user will be presented with the text associated with the content. Each word, with exact reference to its precise positioning within the data stream, will be indexed. Manipulation of text (e.g. cut, copy, paste), will, in effect, be the method of editing that content. Picture and sound will follow along. [Emphasis mine]
Readers of my blog know that I think machine learning is going to revolutionize the edit suite; mainly by reducing the need for Assistant Editors to perform ‘mechanical’ tasks like Ingesting, Sync-ing, and Grouping. But I don’t agree with Ohanian here. And I think his point of view, that editing is basically mechanical, represents one of the problems we face when trying to discuss the future of nonlinear editing.
Editing is a visceral experience. Full stop.
Editing will never be as easy as cutting and pasting text because what’s being said is often secondary to how something’s said. Think about the Brett Kavanaugh hearings. You could read transcripts all day long, but his anger is what left its lasting impact.
The primacy of subtext is applicable to all genres of editing, from the biggest tentpole blockbuster to most corporate HR training video. Anyone who’s listened to multiple reads of Voice Over will know firsthand that the same words spoken differently feel very different each and every time. What makes every editor unique is how these subtle differences inform their creative process.
The source/record metaphor is probably a dated way to interact with audio/video media; and smarter tools that assist the editor in finding and selecting media are needed. But I think “Marking IN and Marking OUT to create edit points” is going to be with us for a while because Marking IN and Marking OUT is editing. The problem isn’t the model, it’s that we need to expand our definition of literacy to include video.