You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
2.**Install the requirements by running this command or following `../docs/installation.md`:**
35
+
```bash
36
+
conda create -y -n lorafusion python=3.12
37
+
conda activate lorafusion
38
+
cd benchmarks_paper
39
+
bash scripts/setup/setup_env.sh
40
+
```
41
+
42
+
3.**Download the Hugging Face models and datasets. Make sure you are logged in and have access to them:**
43
+
```bash
44
+
# huggingface-cli login
45
+
python prepare_models.py
46
+
python gen_sample_distribution.py
47
+
```
48
+
49
+
## Evaluation Workflow
50
+
51
+
### Major Claims
52
+
53
+
-**(C1)**: LoRAFusion is up to 1.96× faster (average 1.47×) than Megatron-LM, and up to 1.46× faster (average 1.29×) than mLoRA. See Section 4.1 and Figure 1.
54
+
55
+
-**(C2)**: Our fused kernels are up to 1.39× faster (average 1.27×) and can replace existing LoRA kernels. See Section 4.2 and Figure 5, Figure 6, and Figure 7.
56
+
57
+
### Experiments
58
+
59
+
1.**Make sure you are in the `benchmarks_paper` directory.**
60
+
61
+
2.**Run the experiments:**
62
+
```bash
63
+
bash scripts/run_all.sh
64
+
```
65
+
66
+
a. This runs all the main experiments and kernel performance tests. It takes about 4 hours.
67
+
68
+
b. Check `scripts/run_all.sh` for the exact commands and timing for each experiment.
69
+
70
+
c. You can easily modify it to run only some experiments.
71
+
72
+
3.**Check the results in the `results` directory. The script automatically creates plots like those in Figure 0, Figure 1, Figure 5, Figure 6, and Figure 7.**
73
+
74
+
## Notes on Reusability
75
+
76
+
To customize experiments, edit `scripts/run_all.sh` and the related sub-scripts.
77
+
We provide detailed scripts for each experiment and corresponding Python scripts to generate the plots.
0 commit comments