diff --git a/README.md b/README.md
index bef3ad58b3de11988d74698c84acb65aadaec8fb..0d18f53ab331c18e334d49477e33ca3cb1736af2 100644
--- a/README.md
+++ b/README.md
@@ -172,9 +172,6 @@ for i, result in enumerate(inference_detector(model, imgs, cfg, device='cuda:0')
 mmdetection implements distributed training and non-distributed training,
 which uses `MMDistributedDataParallel` and `MMDataParallel` respectively.
 
-We suggest using distributed training even on a single machine, which is faster,
-and non-distributed training are left for debugging or other purposes.
-
 ### Distributed training
 
 mmdetection potentially supports multiple launch methods, e.g., PyTorch’s built-in launch utility, slurm and MPI.
@@ -202,6 +199,9 @@ Expected results in WORK_DIR:
 - saved checkpoints (every k epochs, defaults=1)
 - a symbol link to the latest checkpoint
 
+> **Note**
+> 1. We recommend using distributed training with NCCL2 even on a single machine, which is faster. Non-distributed training is for debugging or other purposes.
+> 2. The default learning rate is for 8 GPUs. If you use less or more than 8 GPUs, you need to set the learning rate proportional to the GPU num. E.g., modify lr to 0.01 for 4 GPUs or 0.04 for 16 GPUs.
 
 ## Technical details