Skip to content

Conversation

@sayakpaul
Copy link

What does this PR do?

This PR aims to simplify a few things being discussed in huggingface#12999.

However, I am seeing that in from_pretrained() when use_flashpack=True, a lot of code for dealing with parallelism and quantization are being skipped which is undesirable. Is that expected?

@sayakpaul sayakpaul force-pushed the devanshi00-fal-flashpack branch from c967e88 to ff26d9f Compare January 22, 2026 11:47
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant