8000 HiDream: Unlock max length for some text encoders (t5 and llama) by power88 · Pull Request #7685 · comfyanonymous/ComfyUI · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

HiDream: Unlock max length for some text encoders (t5 and llama) #7685

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

power88
Copy link
Contributor
@power88 power88 commented Apr 20, 2025

Improvement of the last PR: #7676

Currently, T5 and Llama text encoders have restrictive max length limits (77 tokens), which can hinder performance for long-context tasks. This PR removes those constraints, allowing full utilization of the models' capabilities.

Key changes:

  • Unlocks T5xxl limit to 128 tokens
  • Extends Llama limit to 4096 tokens

Note:
The DiT model cannot generate high-quality images despite having long contexts. This may be because DiT wasn't trained for long tokens.
f05085964cd4df218c9c14f5b6bf1593

@power88 power88 requested a review from comfyanonymous as a code owner April 20, 2025 04:33
@comfyanonymous
Copy link
Owner

There is no effective limit to the llama model currently and your 77 limit on the clips is wrong.

@power88
Copy link
Contributor Author
power88 commented Apr 20, 2025

Apologize. I got something wrong.

@power88 power88 closed this Apr 20, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants
0