@bOtM to LocalLlama • 14 days ago2.2x faster at tokens/sec vs rtx 4090 24gb using LLama 3.1 70B-Q4!message-square0arrow-up11arrow-down10file-text
arrow-up11arrow-down1message-square2.2x faster at tokens/sec vs rtx 4090 24gb using LLama 3.1 70B-Q4!@bOtM to LocalLlama • 14 days agomessage-square0file-text