It needs "eyes" for edits that are already baked into uploaded videos. Right now, Underlord is essentially blind to any cuts or edits already present in the source footage.
If the AI could use actual video awareness to analyze the video frames themselves, it would "see" existing edits and be infinitely better able to improve them, and add time remapping, visual improvements like color balancing, and transitions.