You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This caching is for if you are using the same vocabulary as T5. If you want to train a different model, stream the output and save their pretokenized text format, as we do in run_example.py.
So where can add a seqio.CacheDatasetPlaceholder(required=False)?Or is there a more detailed steps in how to use the five original submix you provided, when I've downloaded them.
I noticed the annotation
# If you're using Seqio, we suggest caching your mixture as they take a while to generate.
But I don't know how to do it
The text was updated successfully, but these errors were encountered: