Skip to content

Text in the Dark dataset. The first extremely low light text dataset based on SID and LOL datasets with a total of 3,075 images and 59,975 texts annotated.

Notifications You must be signed in to change notification settings

chunchet-ng/Text-in-the-Dark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 

Repository files navigation

Text in the Dark

Page Views Count

Official repository of the paper: "Text in the Dark: Extremely Low-Light Text Image Enhancement"

Authored By: Chun Chet Ng*, Che-Tsung Lin*, Zhi Qin Tan, Wan Jun Nah, Xinyu Wang, Jie Long Kew, Pohao Hsu, Shang Hong Lai, Chee Seng Chan, Christopher Zach

*Equal Contribution

Released On: December 20, 2024


Project Page | Paper | Text in the Dark Dataset | Bibtex

Abstract:

Extremely low-light text images pose significant challenges for scene text detection. Existing methods enhance these images using low-light image enhancement techniques before text detection. However, they fail to address the importance of low-level features, which are essential for optimal performance in downstream scene text tasks. Further research is also limited by the scarcity of extremely low-light text datasets. To address these limitations, we propose a novel, text-aware extremely low-light image enhancement framework. Our approach first integrates a Text-Aware Copy-Paste (Text-CP) augmentation method as a preprocessing step, followed by a dual-encoder–decoder architecture enhanced with Edge-Aware attention modules. We also introduce text detection and edge reconstruction losses to train the model to generate images with higher text visibility. Additionally, we propose a Supervised Deep Curve Estimation (Supervised-DCE) model for synthesizing extremely low-light images, allowing training on publicly available scene text datasets such as IC15. To further advance this domain, we annotated texts in the extremely low-light See In the Dark (SID) and ordinary LOw-Light (LOL) datasets. The proposed framework is rigorously tested against various traditional and deep learning-based methods on the newly labeled SID-Sony-Text, SID-Fuji-Text, LOL-Text, and synthetic extremely low-light IC15 datasets. Our extensive experiments demonstrate notable improvements in both image enhancement and scene text tasks, showcasing the model’s efficacy in text detection under extremely low-light conditions.

Text in the Dark Dataset

The Text in the Dark dataset is released here.

The Text in the Dark dataset is created based on the combination of the following low light datasets:

  1. See in the Dark (SID) dataset - Sony Set
  2. See in the Dark (SID) dataset - Fuji Set
  3. LOw Light (LOL) dataset

The annotation format for this dataset is following the ICDAR's annotations format:

<x1,y1,x2,y2,x3,y3,x4,y4,text_class> or <L,T,R,B,text_class>, text_class can be "Text" or "###"

Please note that texts of "###" class are ignored during evaluation.

Statistics - Long Exposure Images

SID-Sony Set:

Subset Images Legible Text Illegible Text Total Text
Train 161 5,937 2,128 8,065
Test 50 611 359 970

SID-Fuji Set:

Subset Images Legible Text Illegible Text Total Text
Train 135 6,213 4,534 10,747
Test 41 1,018 1,083 2,101

LOL:

Subset Images Legible Text Illegible Text Total Text
Train 485 613 1,423 2,036
Test 15 28 45 73

Statistics - Short Exposure Images

SID-Sony Set:

Subset Images Legible Text Illegible Text Total Text
Train 280 10,396 3,866 14,262
Test 598 8,210 4,976 13,186

SID-Fuji Set:

Subset Images Legible Text Illegible Text Total Text
Train 286 13,540 10,316 23,856
Test 524 12,768 14,036 26,804

LOL:

Subset Images Legible Text Illegible Text Total Text
Train 485 613 1,423 2,036
Test 15 28 45 73

Citation

If you wish to cite the paper published at ICPR 2022, Extremely Low-Light Image Enhancement with Scene Text Restoration:

@inproceedings{icpr2022_ellie,
  author={Hsu, Po-Hao
  and Lin, Che-Tsung
  and Ng, Chun Chet
  and Long Kew, Jie
  and Tan, Mei Yih
  and Lai, Shang-Hong
  and Chan, Chee Seng
  and Zach, Christopher},
  booktitle={2022 26th International Conference on Pattern Recognition (ICPR)}, 
  title={Extremely Low-Light Image Enhancement with Scene Text Restoration}, 
  year={2022},
  pages={317-323}}

If you wish to cite the latest version of Text in the Dark dataset published at Signal Processing: Image Communication, Text in the Dark: Extremely Low-Light Text Image Enhancement:

@article{LIN2025117222,
  title = {Text in the dark: Extremely low-light text image enhancement},
  journal = {Signal Processing: Image Communication},
  volume = {130},
  pages = {117222},
  year = {2025},
  issn = {0923-5965},
  doi = {https://doi.org/10.1016/j.image.2024.117222},
  url = {https://www.sciencedirect.com/science/article/pii/S0923596524001231},
  author = {Che-Tsung Lin and Chun Chet Ng and Zhi Qin Tan and Wan Jun Nah and Xinyu Wang and 
  Jie Long Kew and Pohao Hsu and Shang Hong Lai and Chee Seng Chan and Christopher Zach},
}

Feedback

We welcome all suggestions and opinions (both positive and negative) on this work. Please contact us by sending an email to ngchunchet95 at gmail.com or cs.chan at um.edu.my.

Acknowledgement

We would like to express our gratitude for the contributions of computing resources and annotation platforms by ViTrox Corporation Berhad. Their generous support has made this work possible.

License and Copyright

This project is open source under the BSD-3 license (refer to the LICENSE file for more information).

© 2024 Universiti Malaya.

About

Text in the Dark dataset. The first extremely low light text dataset based on SID and LOL datasets with a total of 3,075 images and 59,975 texts annotated.

Topics

Resources

Stars

Watchers

Forks