#sqoop 解压软件 $ sudo tar -zxvf sqoop-1.4.6.bin__hadoop-0.23.tar.gz -C /usr/local/ $ pwd /usr/local/sqoop-1.4.6.bin__hadoop-0.23 $ cd conf/ $ sudo cp sqoop-env-template.sh sqoop-env.sh $ sudo vi sqoop-env.sh export HADOOP_COMMON_HOME=/usr/local/hadoop-2.7.4/ export HADOOP_MAPRED_HOME=/usr/local/hadoop-2.7.4/ #sqoop 导入Oracle数据库到HDFS中 $ sqoop import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp --columns 'empno,ename,job,sal,deptno' -m 1 --target-dir '/sqoop/emp' #sqoop 导入Oracle到hive中 $ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno' #导入Oracle,并指明表名 $ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno' --hive-table emp1 #导入Oracle,指定whare条件 $ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno' --hive-table emp2 where 'deptno=10' #导入Oracle,使用select $ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger -m 1 --query 'select * from emp where sal < 1000 and $CONDITIONS' --target-dir '/sqoop/emp3' --hive-table emp3 $ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger -m 1 --table myemp --export-dir '/sqoop/emp3'
免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。