当前位置:网站首页>Instance tunnel use
Instance tunnel use
2022-07-25 15:15:00 【The south wind knows what I mean】
Instance Tunnel Use
1. To perform a select Inquire about :
[email protected] odps_test_project>select * from wc_in;
ID = 20170724071705393ge3csfb8
2. Use Instance Tunnel Download Command to download the execution result to a local file
[email protected] odps_test_project> tunnel download -h true -cf true instance://huorong_bigdata/20220329031048178gumx85t32 C:\Users\HR\Desktop\weekReport\ Weekday Wednesday .csv;\
2017-07-24 15:18:47 - new session: 2017072415184785b6516400090ca8 total lines: 8
2017-07-24 15:18:47 - file [0]: [0, 8), result
downloading 8 records into 1 file
2017-07-24 15:18:47 - file [0] start
2017-07-24 15:18:48 - file [0] OK. total: 44 bytes
download OK
// View results
cat result
slkdfj
hellp
apple
tea
peach
apple
tea
teaa
Parameter description
command :
tunnel download instance://<[project_name/]instance_id> <path>
Be careful
odps The following parameters are set by default
instance_tunnel_max_record=10000
The number of download lines is limited to 1w Within the line , Big watch, go odps_config.ini Put it down in the file , There will be no line limit for downloading
Parameter description
Required parameters
path
Download the path and name of the data file .
There are two options for the archive path of data files : You can archive files directly to MaxCompute Client's bin Directory , here path It needs to be set to the file name . Suffix name ; You can also archive files to other paths , for example D Discoid test Folder ,path I need to set to D:\test\ file name . Suffix name .
[project.]table[/partition]
Target table name . If it's a partition table , You need to specify to the last level partition . If it is not a table in the current space , You need to specify the space name where the table is located .
Optional parameters
-acp
If the specified partition does not exist , Automatically create the target partition . The default value is False.
-bs
Specify each upload to Tunnel Data block size . The default value is 100 MiB(1 MiB=1024×1024 Byte).
-c
Specify local data file encoding . Default not set , Download source data .
-cf
Specify whether to be CSV file , The default value is False.
explain Upload Support only TXT、CSV Format file , Default upload TXT file . If upload CSV file , You need to specify the -cf Parameters , Please also download the latest version of MaxCompute client .
-cp
Specify whether to upload data after local compression , Reduce network traffic . The default value is True.
-dbr
Specify whether to ignore dirty data ( Multiple columns 、 Less column 、 The column data types do not match ). The default value is False.
True: Ignore all data that does not conform to the table definition .
False: If you encounter dirty data , Then an error message will be given , The original data in the target table will not be contaminated .
-dfp
Appoint DATETIME Type data format , The default is yyyy-MM-dd HH:mm:ss. If you want to specify the time format to the millisecond level , have access to tunnel upload -dfp 'yyyy-MM-dd HH:mm:ss.SSS'.DATETIME Please refer to data type version description for details of data type .
-fd
Specify column delimiters for local data files . The default value of comma is English (,).
-h
Specifies whether the data file includes a header . The default value is False. If True, be dship The header will be skipped and the data will be uploaded from the second row .
-mbr
Specify the tolerable amount of dirty data . When this amount of data is exceeded , Termination of the upload . The default value is 1000 strip .
-ni
Appoint NULL Data identifier . The default value is an empty string .
-ow
Specifies whether the uploaded data overwrites the table or partition . The default value is False, The uploaded data is accumulated . Examples are as follows .
tunnel upload -overwrite true log.txt test_project.test_table/p1="b1",p2="b2";
-rd
Specify the row delimiter for the local data file . The default value is \r\n.
-s
Specify whether to scan local data files . The default value is True.
True: Scan the data first , If the data format is correct , Reimport data .
False: Do not scan data , Import data directly .
Only: Scan only local data , Do not continue importing data after the scan .
-sd
Appoint Session Catalog .
-ss
Specify strict schema patterns . The default value is True. If False, The extra data will be discarded , Insufficient fields will be filled in NULL.
-t
Appoint Threads The number of . The default value is 1.
-te
Appoint Tunnel Of Endpoint.
-time
Specify whether to track the upload time . The default value is False.
-tz
Designated time zone . The default value is the local time zone , for example Asia/Shanghai.
边栏推荐
- How to realize a correct double check locking
- Idea远程提交spark任务到yarn集群
- Use the command to check the WiFi connection password under win10 system
- NPM's nexus private server e401 E500 error handling record
- Award winning interaction | 7.19 database upgrade plan practical Summit: industry leaders gather, why do they come?
- 什么是物联网
- API health status self inspection
- Scala110-combineByKey
- Vscode plugin collection
- 万能通用智能JS表单验证
猜你喜欢

bridge-nf-call-ip6tables is an unknown key异常处理

Recommend 10 learning websites that can be called artifact

"How to use" decorator mode

Implement a simple restful API server

Share a department design method that avoids recursion
![[Nacos] what does nacosclient do during service registration](/img/76/3c2e8f9ba19e36d9581f34fda65923.png)
[Nacos] what does nacosclient do during service registration

ice 100G 网卡分片报文 hash 问题

SPI传输出现数据与时钟不匹配延后问题分析与解决

String type time comparison method with error string.compareto

Spark提交参数--files的使用
随机推荐
Vscode plugin collection
Leo-sam: tightly coupled laser inertial odometer with smoothing and mapping
Client error: invalid param endpoint is blank
Handle Oracle deadlock
"How to use" decorator mode
Visual Studio 2022 查看类关系图
[comprehensive pen test] difficulty 4/5, classic application of line segment tree for character processing
Idea远程提交spark任务到yarn集群
处理ORACLE死锁
vscode 插件篇收集
Run redis on docker to start in the form of configuration file, and the connection client reports an error: server closed the connection
什么是物联网
我的创作纪念日
Detailed explanation of lio-sam operation process and code
TypeScript学习2——接口
dpdk 收发包问题案例:使用不匹配的收发包函数触发的不收包问题定位
npm的nexus私服 E401 E500错误处理记录
Gbdt source code analysis of boosting
System. Accessviolationexception: an attempt was made to read or write to protected memory. This usually indicates that other memory is corrupted
"How to use" observer mode