Skip to content

Always Prompting Out of Memory Error (whole all Unable to allocate)

If you're running out of VRAM, try these solutions. The "out of memory" error indicates that your GPU doesn't have enough memory to perform the current operation. To resolve this, consider the following:

  1. Use Smaller Models: Opt for smaller models like small, medium, or base instead of large-v3.
  2. Segmentation Options: If you still want to use a large model, choose "Pre-Segmentation" or "Equal Segmentation".
  3. Advanced Settings Adjustments: Modify the advanced options in the menu bar under Tools/Options:
CUDA data type = float32 to int8. If you encounter errors, change it to float16.
beam_size = 5 to 1
best_of = 5 to 1
Context = true to false