How to copy data between s3 buckets
Overview Transferring data between Amazon S3 buckets is a common requirement for many AWS users. This guide will walk …
In this article we will go over some frequently used HDFS commands
The Hadoop Distributed File System (HDFS) is a distributed file system designed to run on commodity hardware. It has many similarities with existing distributed file systems. However, the differences from other distributed file systems are significant. HDFS is highly fault-tolerant and is designed to be deployed on low-cost hardware. HDFS provides high throughput access to application data and is suitable for applications that have large data sets. HDFS commands helps to perform Hadoop HDFS file operations.
Command : hadoop version
Command: hadoop fs -ls
Command: hadoop fs -df hdfs
Command: hadoop fs -df -h hdfs
Command: hadoop fs -mkdir /user/entechlog/hadoop
Command:hadoop fs -put /data/home/entechlog/hadoop/input/HadoopFile1.txt /user/entechlog/hadoop/input
Command: hadoop fs -copyFromLocal /data/home/entechlog/hadoop/input/HadoopFile1.txt /user/entechlog/hadoop/input/
Command: hadoop fs -ls /user/entechlog/hadoop/input/
Command:hadoop fs -put /data/home/entechlog/hadoop/input/copydir/ /user/entechlog/hadoop/input/
Command: hadoop fs -du -s -h /user/entechlog/hadoop/input/copydir/
Command: hadoop fs -rm /user/entechlog/hadoop/input/copydir/HadoopFile1.txt
Command: hadoop fs -ls /user/entechlog/hadoop/input/copydir/HadoopFile1.txt
Command: hadoop fs -rm /user/entechlog/hadoop/input/copydir/\*
Command: hadoop fs –expunge
Command: hadoop fs -rm -r /user/entechlog/hadoop/input/copydir/
Command: hadoop fs -cp /user/entechlog/hadoop/input/HadoopFile3.txt /user/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -copyToLocal /user/entechlog/hadoop/input/HadoopFile4.txt /data/home/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -get /user/entechlog/hadoop/input/HadoopFile4.txt /data/home/entechlog/hadoop/input/HadoopFile5.txt
Command: hadoop fs -tail /user/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -chmod 600 /user/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -chown newUser:newGroup /user/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -chgrp newGroup /user/entechlog/hadoop/input/HadoopFile4.txt
Command: hadoop fs -mv source target
Command: hadoop fs
Command: hadoop fs –help
Command: hadoop dfsadmin -safemode get
Command: hadoop dfsadmin -safemode enter
Command: hadoop dfsadmin -safemode leave
Overview Transferring data between Amazon S3 buckets is a common requirement for many AWS users. This guide will walk …
In this article, we’re diving into a data engineering demo project that’s all about bringing cricket player …