Upload new Sqoop tarballs to HDFS, if you have not already done so in the prior Hive and WebHCat step. Execute these commands on the host with the Sqoop client:
su -l <HDFS_USER> -c "hdfs dfs -mkdir -p /hdp/apps/2.3.x.y-z/sqoop/" su -l <HDFS_USER> -c "hdfs dfs -put /usr/hdp/current/sqoop-client/sqoop.tar.gz /hdp/apps/2.3.x.y-z/sqoop/" su -l <HDFS_USER> -c "hdfs dfs -chmod -R 555 /hdp/apps/2.3.x.y-z/sqoop" su -l <HDFS_USER> -c "hdfs dfs -chmod -R 444 /hdp/apps/2.3.x.y-z/sqoop/sqoop.tar.gz" su -l <HDFS_USER> -c "hdfs dfs -chown -R <HDFS_USER>:<HADOOP_GROUP>/hdp"
where <HDFS_USER> is the HDFS Service user. For example, hdfs.
Save the old Sqoop configuration and add symlink from /etc/sqoop/conf:
mv /etc/sqoop/conf /etc/sqoop/conf.saved
ln -s /usr/hdp/current/sqoop-client/conf /etc/pig/conf
ls -la /etc/sqoop
total 4 drwxr-xr-x 3 root root 4096 Jun 19 21:51 2.3.0.0-2323 lrwxrwxrwx 1 root root 35 Jun 19 21:54 conf -> /usr/hdp/current/sqoop-client/conf drwxr-xr-x 2 root root 4096 Jun 14 00:11 conf.saved