FeepingCreature f90ec0e6e2 Add --use-flash-attention flag.
This is useful on AMD systems, as FA builds are still 10% faster than Pytorch cross-attention.
2025-03-13 23:27:06 +01:00
..
2024-12-20 16:24:55 -05:00
2025-01-16 17:45:37 -05:00
2024-12-12 17:59:16 -05:00
2024-12-16 19:35:40 -05:00
2025-03-13 23:27:06 +01:00
2024-12-27 18:02:30 -05:00
2025-02-27 07:22:42 -05:00