I am looking forward to latent coordinates plus model being metadata for some frames of videos at least.
You don’t need total precision for every visual representation but it could work as a great compression technique. Assuming we get the GenAI power usage down.
I personally would love to see better simulation of complex systems in games. Games are how we as humans explore the world in safe constraints to learn and grow with less risk. A lot of the limits of games though are just limits of the creators understanding and level of effort it takes to represent that detail of the world, but it means that lesson around the now missing detail can’t be learned.
Another one for me, tailored voice and visuals for technical talks.
Again a lot of what is trying to convened is the actual technical content, but language, accents, verbal tics, cultural specific metaphors, generic or uninteresting visuals can all act as a barrier to that information. Seeing automatic content translation to improve my personal viewing style would be awesome to me!
Have you messed with RAG yet? That the leg in the journey to me. I am hoping it will help a little with the “sketchy” part of info.