From 2e6b8d0a218c2a5cb67335aae473271fd14483e7 Mon Sep 17 00:00:00 2001 From: Fred <59241275+pengzhangzhi@users.noreply.github.com> Date: Tue, 3 Dec 2024 09:10:14 -0500 Subject: [PATCH] Update README.md --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 0b312f6..859a258 100644 --- a/README.md +++ b/README.md @@ -80,8 +80,8 @@ It's recommended to use the flash attention for training. Because in the forward # Benchmarking -### FAESM2 vs. Official ESM2 -Below is the comparison of peak memory usage and inference time of FAESM with the official ESM2 and shows that FAESM can save memory usage by up to 60% and inference time by up to 70% (length 1000). The benchmarking is done on ESM-650M with batch size 8, and a single A100 with 80GB of memory. +### FAESM vs. Official ESM2 +Below is the comparison of peak memory usage and inference time of FAESM with the official ESM2. We show that FAESM can save memory usage by up to 60% and inference time by up to 70% (length 1000). The benchmarking is done on ESM-650M with batch size 8, and a single A100 with 80GB of memory. ![benchmark](assets/figs/benchmark.png)