Breaking Limits in Scene Text Recognition: CUEE MDAP Lab’s AI Super-Resolution Model

From the CUEE MDAP Lab (Multimedia Data Analytics and Processing Research Unit), Department of Electrical Engineering, Chulalongkorn University

Turning Blurry Words into Crystal-Clear Text with AI

Have you ever taken a photo of a sign or document, only to find the text too blurry to read? This is a big challenge not just for us, but also for computers. When machines try to read text from low-quality images—whether it’s for traffic signs, street names, or scanned documents—they often struggle. This problem is known as Scene Text Recognition (STR).

To fix this, researchers developed a technique called Scene Text Image Super-Resolution (STISR). Think of it like giving glasses to a blurry image: it sharpens the text, making it clearer and easier for both humans and AI systems to understand.

An illustration comparing blurry and clear text recognition using AI. On the left, labeled “Before (LR),” a hand holds a smartphone showing distorted letters “W*??R.” On the right, labeled “After (STISR),” the word “WATER” appears clear through glasses, representing AI Super-Resolution technology. The image symbolizes how CUEE MDAP Lab’s AI model transforms unreadable scene text into sharp, understandable words.

📸 The Problem with Blurry Text

Cameras in real life—like the ones in phones, drones, or surveillance systems—don’t always capture perfect images. Poor lighting, shaky hands, or distance can make text hard to read. Traditional methods to sharpen images improve the picture, but they often miss the fine details that matter most for text: edges, strokes, and character shapes.

A comparative image showing several low-resolution scene text samples labeled “LR,” each displaying blurry or partially unreadable words such as “wator,” “vro,” “lons,” and “cllm.” The image demonstrates the visual challenge of recognizing text from unclear, low-quality images — a problem addressed by CUEE MDAP Lab’s AI Super-Resolution Model for accurate scene text recognition.
Figure 1 Example of a low-resolution scene text image

🤖 Enter MADN: Multi-Attention with Diffusion Network

A new model called MADN has been designed to solve this exact issue. Instead of just brightening or sharpening the whole image, MADN uses smart AI techniques to focus only on the important parts—the letters themselves.

Here’s how it works:

  • Attention Modules: Like a human eye focusing on key details, MADN has a system that decides what parts of the image matter most. It zooms in on the letters while ignoring the background.
  • Sequential Learning: Text isn’t just shapes; it’s a sequence of letters. MADN uses a memory mechanism (BLSTM) to understand how characters relate to each other, helping it guess missing or blurry parts more accurately.
  • Diffusion Process: This clever trick gradually removes noise and sharpens details, almost like a digital artist carefully redrawing each letter.

The result? Text images that go from unreadable blobs to clear, structured words.

A comparison chart showing the improvement from low-resolution text images (labeled “LR”) to enhanced results produced by MADN (Ours). The top row displays blurry, unreadable words like “wator” and “cllm,” while the bottom row shows AI-restored text such as “water” and “children.” The green arrow highlights how CUEE MDAP Lab’s MADN model transforms unclear text into clear, readable characters through advanced AI Super-Resolution technology.
Figure 2 MADN restores blurry text into clear, readable words

In tests with the TextZoom dataset (a benchmark for this problem), MADN outperformed previous models, delivering sharper images and higher recognition accuracy.

A comparison chart showing text restoration results across multiple AI models on the TextZoom dataset. Rows represent methods such as Bicubic, SRCNN, SRResNet, RDN, TSRN, and MADN (Ours), while columns show Easy, Medium, and Hard subsets with sample words like “regular,” “work,” “THURSDAYS,” and “POLYTECHNIC.” The MADN model, highlighted in a red box, produces the clearest and most accurate text outputs, demonstrating superior performance in scene text super-resolution and recognition.
Figure 3 MADN outperforms existing models on TextZoom dataset

🚀 A Step Toward Smarter Vision

What makes MADN stand out is its balance of accuracy and efficiency—delivering state-of-the-art text clarity with moderate compute. The paper also notes real-time optimization for edge deployments in autonomous and smart-city scenarios as the next step.


🌍 Why This Matters

This technology isn’t just academic. It has a real-world impact in:

A futuristic autonomous car driving in rainy, foggy weather illuminated by streetlights, with a glowing AI-enhanced road sign reading “STOP.” The image represents how CUEE MDAP Lab’s AI Super-Resolution Model enables vehicles to recognize road signs clearly in poor weather conditions for safer autonomous driving.

Autonomous Driving

Reading road signs in poor weather.

A futuristic smart city street at night showing an old, faded road sign digitally restored by AI. The left half of the sign appears cracked and weathered, while the right half glows bright and sharp with cyan holographic effects — symbolizing how CUEE MDAP Lab’s AI Super-Resolution Model brings clarity and modernization to city infrastructure.

Smart Cities

Digitizing old or damaged signs.

A futuristic document scanning scene showing a paper split into two halves — one side blurry and faded, the other sharp and clear after AI enhancement. A glowing cyan scanning beam runs across the page, representing how CUEE MDAP Lab’s AI Super-Resolution Model restores lost details and improves text clarity in digital document processing.

Document Scanning

Recovering details from low-quality scans.

A futuristic surveillance scene divided into two halves — the left side shows blurry, low-quality CCTV footage of a street and vehicle, while the right side is enhanced with crystal-clear detail and a visible license plate. Cyan holographic overlays symbolize how CUEE MDAP Lab’s AI Super-Resolution Model clarifies camera footage for advanced security and surveillance analysis.

Security and Surveillance

Making sense of blurry camera footage.


In Short: The MADN model gives AI-powered “glasses” to blurry text images, making them readable and useful again. It’s a powerful step forward in how machines—and people—can interact with the world through clearer, sharper digital vision.


🔗 Learn more & official links

📬 Contact Us

Interested in research from the CUEE MDAP Lab—including this MADN work or other lab projects such as:

  • Anomaly Detection
  • Digital Image and Video Super‑resolution Techniques
  • Digital Video Coding
  • Face Recognition and Emotional Expression

Please reach out via Design Gateway to connect with the team and learn more.

👉 Contact Design Gateway