[root@hadoop01 ~]# lltotal 156drwxr-xr-x. 18 root root 4096 Jan 5 05:05 apps-rw-r--r--. 1 root root 156 Jul 24 21:47 common-friends.logdrwxr-xr-x. 2 root root 4096 Oct 9 09:18 datadrwxr-xr-x. 2 root root 4096 Aug 29 01:52 esdata-rw-r--r--. 1 root root 2231 Jul 24 22:00 flow.datdrwxr-xr-x. 9 root root 4096 Oct 8 08:13 flumedrwxr-xr-x. 2 root root 4096 Oct 30 18:30 hadooptestdrwxr-xr-x. 5 root root 4096 Oct 4 15:12 hdpdatadrwxr-xr-x. 4 root root 4096 Oct 5 10:17 hivedatadrwxr-xr-x. 3 root root 4096 Sep 25 13:59 journaldatadrwxr-xr-x. 5 root root 4096 Oct 13 13:55 logsdrwxr-xr-x. 2 root root 4096 Oct 22 09:46 out123-rwxr-xr-x. 1 root root 4482 Aug 19 09:57 PWDdrwxr-xr-x. 3 root root 4096 Dec 21 07:02 softdrwxr-xr-x. 12 hadoop hadoop 4096 Jun 21 2016 spark-1.6.2-bin-hadoop2.6drwxr-xr-x. 9 root root 4096 Oct 10 10:23 sqoopdrwxr-xr-x. 5 root root 4096 Nov 17 08:17 storm-local-rw-r--r--. 1 root root 660 Aug 28 06:33 storm.yaml-rw-r--r--. 1 root root 124 Sep 28 10:06 user_prefer.sh-rw-r--r--. 1 root root 7495 Nov 16 03:42 wordcount.jar-rw-r--r--. 1 root root 56 Oct 13 16:46 word.txtdrwxr-xr-x. 3 root root 4096 Jan 5 18:49 zkdata-rw-r--r--. 1 root root 54453 Jan 5 18:49 zookeeper.out[root@hadoop01 ~]# du -sh . [对当前目录下所有的目录和文件的大小进行汇总,-s表示汇总,-h表示以KB, MB, GB, TB格式进行人性化显示]6.6G .[root@hadoop01 ~]# du -sh * [对当前目录下每一个目录和文件的大小分别进行汇总]3.2G apps4.0K common-friends.log416K data4.0K esdata4.0K flow.dat82M flume4.0K hadooptest150M hdpdata92M hivedata66M journaldata344M logs16K out1238.0K PWD2.2G soft303M spark-1.6.2-bin-hadoop2.627M sqoop52K storm-local4.0K storm.yaml4.0K user_prefer.sh8.0K wordcount.jar4.0K word.txt260M zkdata60K zookeeper.out[root@hadoop01 ~]# du -sh apps [对指定的目录的大小进行汇总]3.2G apps[root@hadoop01 ~]# du -sh logs/* [对指定的目录下面所有的子目录和文件的大小分别进行汇总]39M logs/click_log8.0K logs/es-logs306M logs/kafka-logs380K logs/test.log[root@hadoop01 ~]# du -sh *|grep [M] [对当前目录下所有的目录和文件分别进行汇总,并使用"grep [M]"筛选出所有大小为M(兆)级别的目录和文件]82M flume150M hdpdata92M hivedata66M journaldata344M logs303M spark-1.6.2-bin-hadoop2.627M sqoop260M zkdata[root@hadoop01 ~]# du -sh *|grep [G] [筛选出大小为GB级别的目录和文件]3.2G apps2.2G soft[root@hadoop01 ~]# du -sh *|grep [MG] [筛选出大小为MB和GB级别的目录和文件]3.2G apps82M flume150M hdpdata92M hivedata66M journaldata344M logs2.2G soft303M spark-1.6.2-bin-hadoop2.627M sqoop260M zkdata[root@hadoop01 ~]# du -sh *|grep [MG]|sort -nr [筛选出大小为MB和GB级别的目录和文件,并降序排序]344M logs303M spark-1.6.2-bin-hadoop2.6260M zkdata150M hdpdata92M hivedata82M flume66M journaldata27M sqoop3.2G apps2.2G soft[root@hadoop01 ~]#
如果觉得本文对您有帮助,不妨扫描下方微信二维码打赏点,您的鼓励是我前进最大的动力: