DeepSeek-R1¶
The distilled version of Deepseek-R1 models are now supported for both performance benchmarking and model evaluations 🎉. You can use built in support for 4 different datasets: LongBench
, Dolly
, OpenOrca
and ConvFinQA
. You can deploy the Deepseek-R1 distilled models on Amazon EC2, Amazon Bedrock or Amazon SageMaker.
The easiest way to benchmark the DeepSeek models is through the FMBench-orchestrator
on Amazon EC2 VMs.
Benchmark Deepseek-R1 distilled models on Amazon EC2¶
👉 Make sure your account has enough service quota for vCPUs to run this benchmark. We would be using g6e.xlarge
, g6e.2xlarge
, g6e.12xlarge
and g6e.48xlarge
instances, if you do not have sufficient service quota then you can set deploy: no
in the configs/deepseek/deepseek-convfinqa.yml
(or other) file to disable some tests as needed.
Follow instructions here to install the orchestrator. Once installed you can run Deepseek-r1 benchmarking with the ConvFinQA
dataset the following command:
--config-file
parameter to configs/deepseek/deepseek-longbench.yml
or configs/deepseek/deepseek-openorca.yml
to use other datasets for benchmarking. These orchestrator files test various Deepseek-R1 distilled models on g6e
instances, edit this file as per your requirements.
Benchmark Deepseek-R1 quantized models on Amazon EC2¶
👉 Make sure your account has enough service quota for vCPUs to run this benchmark. We would be using g6e.12xlarge
instance for this test.
-
Create a
g6e.12xlarge
instance and run theDeepSeek-R1 1.58b quantized
model on this instance by following the steps 1 through 8 described here. -
Follow steps 1 through 5 here to setup
FMBench
on this instance. -
Next run the following command to benchmark LongBench
-
Once the run completes you should see the benchmarking results in a folder called
results-DeepSeek-R1-quant-1.58bit-g6e.12xl
present in your current directory.