Precomputing large feature vectors #4458
Replies: 1 comment
-
Hi! You can reduce dset_with_embeddings = dset.map(lambda ex: {"embedding": compute_embedding(ex)}, writer_batch_size=10) |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I want to precompute a feature vector of 1900 dimensions for over 500 data points and store it in an HF dataset. RAM is becoming a problem. How can I make sure to write directly to disk?
Beta Was this translation helpful? Give feedback.
All reactions