-
Notifications
You must be signed in to change notification settings - Fork 4.4k
fix(extraction): llm extraction strategy not overriding response #1150
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
If `response` is overriden, and some error is met in later part of this `try`, in `except` we'll get error triggering response.choices[0].message.content.
WalkthroughThe variable used to store the LLM response content in the Changes
Poem
Note ⚡️ AI Code Reviews for VS Code, Cursor, WindsurfCodeRabbit now has a plugin for VS Code, Cursor and Windsurf. This brings AI code reviews directly in the code editor. Each commit is reviewed immediately, finding bugs before the PR is raised. Seamless context handoff to your AI code agent ensures that you can easily incorporate review feedback. Warning Review ran into problems🔥 ProblemsGit: Failed to clone repository. Please run the 📜 Recent review detailsConfiguration used: CodeRabbit UI 📒 Files selected for processing (1)
🔇 Additional comments (3)
✨ Finishing Touches
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Create a ticket on our support page for assistance with any issues or questions. Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
CodeRabbit Configuration File (
|
Hi @unclecode @aravindkarnam I meet a problem of error |
Summary
In extract() of LLMExtractionStrategy,
response
is not overridden. Thus if some error is met in later part of thistry
, inexcept
we'll get error triggering response.choices[0].message.content:'str' object has no attribute 'choices'
, as response is now overridden into a string.https://github.com/unclecode/crawl4ai/blob/main/crawl4ai/extraction_strategy.py#L658-L689
This error occurs occasionally when crawl4ai is used against a list of models include gemini-2.5-pro and gemini-2.5-flash.
List of files changed and why
In extract() of LLMExtractionStrategy, change
response = response.choices[0].message.content
tocontent = ...
, so response is not overriden.How Has This Been Tested?
Extract for 200 different webpages after fixing this on my computer.
Checklist:
Summary by CodeRabbit