mongoexport

导出csv格式:默认为json格式

1,导出csv格式数据,需要同时指定:--type=csv     --fields  column_a,column_b,column_c 

mongoexport --authenticationDatabase admin   -hzzz  --port=28000 -uxxx -pyyy -d apple   -c users   --type=csv --fields _id,username,domain,name,display_name,department_name  -o  users.csv

2,导出csv格式数据,需要同时指定:--type=csv   --fieldFile

[work@hostname tmp]$ cat fields.txt
_id
username
domain
[work@hostname tmp]$ 

$ mongoexport --authenticationDatabase admin   -hzzz  --port=28000 -uxxx -pyyy -d apple   -c users   --type=csv --fieldFile  fields.txt  -o  users.csv


#####################

这里就会按照_id,username,domain三个字段来导出:
_id,username,domain
ObjectId(5db805516e1c8355a15fe80a),glc,google.com
ObjectId(5db805516e1c8355a15fe80b),wjl,apple.com
ObjectId(5db805516e1c8355a15fe80c),zd,airbnb.com
ObjectId(5db805516e1c8355a15fe80d),lt,amazon.com

3,导出csv格式数据文件可以不要列名称 ,加上 --noHeaderLine就能让导出得文件不包含: _id,username,domain ,常用:

[work@hostname tmp]$ cat fields.txt
_id
username
domain
[work@hostname tmp]$ 

$ mongoexport --authenticationDatabase admin   -hzzz  --port=28000 -uxxx -pyyy -d apple   -c users   --type=csv --fieldFile  fields.txt --noHeaderLine -o  users.csv


#####################

这里就会按照_id,username,domain三个字段来导出:

ObjectId(5db805516e1c8355a15fe80a),glc,google.com
ObjectId(5db805516e1c8355a15fe80b),wjl,apple.com
ObjectId(5db805516e1c8355a15fe80c),zd,airbnb.com
ObjectId(5db805516e1c8355a15fe80d),lt,amazon.com

 按照指定条件导出数据:--query   '{}'

将名称为“glc”的全部导出:

[work@hostname tmp]$mongoexport --authenticationDatabase admin -hxxx  --port=28000 -uyyy -pzzz -d apple -c users --type=csv --fieldFile fields.txt --query '{"username":"glc"}' -o users.csv.glc
2020-10-30T16:57:00.563+0800    connected to: xxx:28000
2020-10-30T16:57:00.743+0800    exported 10 records
[work@hostanme tmp]$ less cas_users.csv.glc
_id,username,domain
ObjectId(7db805916e5c8355a35fe80a),glc,apple.com
ObjectId(7e05a1456e5c836c0c0a1240),glc,apple.com
ObjectId(7e05a3f66e5c836c0c0ac235),glc,apple.cn
ObjectId(7e05a53a6e5c836c0c0b19c4),glc,google.cn
ObjectId(7e05a5ee6e5c836c0c0b5d2d),glc,google.cn
ObjectId(7e05a77d6e5c836c0c0b8fb8),glc,google.cn
ObjectId(7e3920b6d3dbdb333aa04844),glc,google.cn
ObjectId(7e4e625ed3dbdb333aa04f7e),glc,amazon.com
ObjectId(7e96c569e566e174015b7fae),glc,amazon.com
ObjectId(7f4c80abd3dbdb3d702f7a00),glc,apple.com
[work@hostname tmp]$ 

 ########################################################

mongoexport导出某个集合:

/home/work/mongodb/4.0/bin/mongoexport --authenticationDatabase admin --host 10.10.10.10 --port 28000 --username mongo_backup --password 123456  --db db_name --collection table_name --out /home/work/tmp/table_name

 从上面的结果可以看出,我们在导出数据时没有显示指定导出样式 ,默认导出了JSON格式的数据。

mongoexport --host hostname --port 28000 --username user --authenticationDatabase admin --collection table_name --db db_name --out table_name.json
mongoexport --uri 'mongodb://user:password@hostname:28000/db_name?authsource=admin' --collection table_name --out table_name.json
mongoexport --db db_name --collection table_name --query '{"dept": "ABC", date: { $gte: { "$date": "2018-01-01T00:00:00.000Z" } }}'

uri规则:

--uri "mongodb://[username:password@]host1[:port1][,host2[:port2],...[,hostN[:portN]]][/[database][?options]]"
[work@hostname tmp]$ mongoimport  --help
Usage:
  mongoimport <options> <file>

Import CSV, TSV or JSON data into MongoDB. If no file is provided, mongoimport reads from stdin.

See http://docs.mongodb.org/manual/reference/program/mongoimport/ for more information.

general options:
      --help                                      print usage
      --version                                   print the tool version and exit

verbosity options:
  -v, --verbose=<level>                           more detailed log output (include multiple times for more verbosity, e.g. -vvvvv, or specify a
                                                  numeric value, e.g. --verbose=N)
      --quiet                                     hide all log output

connection options:
  -h, --host=<hostname>                           mongodb host to connect to (setname/host1,host2 for replica sets)
      --port=<port>                               server port (can also use --host hostname:port)

kerberos options:
      --gssapiServiceName=<service-name>          service name to use when authenticating using GSSAPI/Kerberos ('mongodb' by default)
      --gssapiHostName=<host-name>                hostname to use when authenticating using GSSAPI/Kerberos (remote server's address by default)

ssl options:
      --ssl                                       connect to a mongod or mongos that has ssl enabled
      --sslCAFile=<filename>                      the .pem file containing the root certificate chain from the certificate authority
      --sslPEMKeyFile=<filename>                  the .pem file containing the certificate and key
      --sslPEMKeyPassword=<password>              the password to decrypt the sslPEMKeyFile, if necessary
      --sslCRLFile=<filename>                     the .pem file containing the certificate revocation list
      --sslAllowInvalidCertificates               bypass the validation for server certificates
      --sslAllowInvalidHostnames                  bypass the validation for server name
      --sslFIPSMode                               use FIPS mode of the installed openssl library

authentication options:
  -u, --username=<username>                       username for authentication
  -p, --password=<password>                       password for authentication
      --authenticationDatabase=<database-name>    database that holds the user's credentials
      --authenticationMechanism=<mechanism>       authentication mechanism to use

namespace options:
  -d, --db=<database-name>                        database to use
  -c, --collection=<collection-name>              collection to use

uri options:
      --uri=mongodb-uri                           mongodb uri connection string

input options:
  -f, --fields=<field>[,<field>]*                 comma separated list of fields, e.g. -f name,age
      --fieldFile=<filename>                      file with field names - 1 per line
      --file=<filename>                           file to import from; if not specified, stdin is used
      --headerline                                use first line in input source as the field list (CSV and TSV only)
      --jsonArray                                 treat input source as a JSON array
      --parseGrace=<grace>                        controls behavior when type coercion fails - one of: autoCast, skipField, skipRow, stop
                                                  (defaults to 'stop') (default: stop)
      --type=<type>                               input format to import: json, csv, or tsv (defaults to 'json') (default: json)
      --columnsHaveTypes                          indicated that the field list (from --fields, --fieldsFile, or --headerline) specifies types;
                                                  They must be in the form of '<colName>.<type>(<arg>)'. The type can be one of: auto, binary,
                                                  bool, date, date_go, date_ms, date_oracle, double, int32, int64, string. For each of the date
                                                  types, the argument is a datetime layout string. For the binary type, the argument can be one
                                                  of: base32, base64, hex. All other types take an empty argument. Only valid for CSV and TSV
                                                  imports. e.g. zipcode.string(), thumbnail.binary(base64)

ingest options:
      --drop                                      drop collection before inserting documents
      --ignoreBlanks                              ignore fields with empty values in CSV and TSV
      --maintainInsertionOrder                    insert documents in the order of their appearance in the input source
  -j, --numInsertionWorkers=<number>              number of insert operations to run concurrently (defaults to 1) (default: 1)
      --stopOnError                               stop importing at first insert/upsert error
      --mode=[insert|upsert|merge]                insert: insert only. upsert: insert or replace existing documents. merge: insert or modify
                                                  existing documents. defaults to insert
      --upsertFields=<field>[,<field>]*           comma-separated fields for the query part when --mode is set to upsert or merge
      --writeConcern=<write-concern-specifier>    write concern options e.g. --writeConcern majority, --writeConcern '{w: 3, wtimeout: 500, fsync:
                                                  true, j: true}'
      --bypassDocumentValidation                  bypass document validation

#################

###############################

原文地址:https://www.cnblogs.com/igoodful/p/13902621.html