I’ve noticed that although we have a “Preserve Text” function, the results can be a bit off, especially with low-resolution images. While we humans can still make out the text, machines don’t always get it right. I was thinking, what if we could select specific pixels, identify the language, and manually input the text we read? It might just help the software enhance text clarity.