Skip to content

Conversation

@lybtt
Copy link
Contributor

@lybtt lybtt commented Oct 20, 2025

No description provided.

@siyuan-youtu
Copy link
Contributor

Thank you for your attention and contribution.

For YouTube-GraphRAG, during the graph construction, since it inherently requires obtaining the complete LLM output for processing, switching to a streaming approach would not be very meaningful. During the inference phase, the primary factors affecting the speed and user experience are the question decomposition and the multiple retrieval iterations, where the time cost is dominated by the LLM's output generation.

Since these are intermediate steps, switching to streaming output is probably unnecessary. Only the final step, where the LLM generates the answer based on the retrieved context, would be meaningful to change to streaming output.

Additionally, if there are any files unrelated to this specific change, it is recommended to remove them from the commit.

@lybtt
Copy link
Contributor Author

lybtt commented Oct 21, 2025

Thank you for your attention and contribution.

For YouTube-GraphRAG, during the graph construction, since it inherently requires obtaining the complete LLM output for processing, switching to a streaming approach would not be very meaningful. During the inference phase, the primary factors affecting the speed and user experience are the question decomposition and the multiple retrieval iterations, where the time cost is dominated by the LLM's output generation.

Since these are intermediate steps, switching to streaming output is probably unnecessary. Only the final step, where the LLM generates the answer based on the retrieved context, would be meaningful to change to streaming output.

Additionally, if there are any files unrelated to this specific change, it is recommended to remove them from the commit.

Just wanted to add that for lengthy text outputs, especially with privately deployed models that often have shorter timeout settings, streaming can help avoid timeouts by delivering content incrementally.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants