-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
From Authors: Regarding how to test your own data #17
Comments
Thank you for the update and for your hard work! We completely understand that with the CVPR deadline approaching, it's a particularly busy time for everyone. |
Have you tested this on any ocr intensive images ? I was very curious of whether you could handle text degradation usually seen in scanned images of bills/textbooks |
Hi there, I have successfully implemented the functionality to test custom data by adding the extraction process for text prompts and T5 features. This now allows for flexible usage beyond the current super-resolution from 256 to 1024. Here are the steps to achieve this: Step 1: Use the MLLM LLaVa to generate a description of the image and obtain the caption.
Command:
Step 2: Use tools/extract_t5_features.py to extract T5 features, resulting in
Step 3: Image Restoration
If anyone needs help or details on this implementation, please feel free to reach out! |
But with your script I got square output and not properly 4:3 ratio. |
We have released more user-friendly inference code. Feel free to test your own images! |
Currently, the inference code in the codebase has not yet been updated to include the extraction process for text prompts and T5 features. The current code only supports super-resolution from 256 to 1024, and support for arbitrary resolutions has not yet been added. I will update these parts as soon as possible; it's a busy time with the CVPR deadline approaching. Thank you all for your patience and understanding! I apologize for any inconvenience this may have caused you.
Updates
We have released more user-friendly inference code. Feel free to test your own images!
The text was updated successfully, but these errors were encountered: