1.使用datax同步数据

2.模板json(已配置hdfs ha):

{
    "job": {
        "content": [
            {
                "reader": {
                    "name": "postgresqlreader",
                    "parameter": {
                        "username": "hs_sync",
                        "password": "Pass2025",
                        "column": [
                            "order_date",
                            "day",
                            "iso_day_of_week",
                            "weekday_cn",
                            "weekday_en",
                            "weekday_short",
                            "is_weekend",
                            "iso_week",
                            "month",
                            "month_cn",
                            "month_en_full",
                            "month_en_short",
                            "quarter",
                            "year"
                        ],
                        "connection": [
                            {
                                "table": [
                                    "dim_calendar"
                                ],
                                "jdbcUrl": [
                                    "jdbc:postgresql://100.64.0.10:25432/hs_sync_data"
                                ]
                            }
                        ],
                        "fetchSize": 1000
                    }
                },
                "writer": {
                    "name": "hdfswriter",
                    "parameter": {
                        "defaultFS": "hdfs://nameservice1",
                        "hadoopConfig": {
                            "dfs.nameservices": "nameservice1",
                            "dfs.ha.namenodes.nameservice1": "nn1,nn2",
                            "dfs.namenode.rpc-address.nameservice1.nn1": "192.168.33.61:8020",
                            "dfs.namenode.rpc-address.nameservice1.nn2": "192.168.33.62:8020",
                            "dfs.client.failover.proxy.provider.nameservice1": "org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
                        },
                        "fileType": "orc",
                        "path": "/tmp/data/test/",
                        "fileName": "dim_calendar",
                        "writeMode": "truncate",
                        "column": [
                            {
                                "name": "order_date",
                                "type": "date"
                            },
                            {
                                "name": "day",
                                "type": "smallint"
                            },
                            {
                                "name": "iso_day_of_week",
                                "type": "smallint"
                            },
                            {
                                "name": "weekday_cn",
                                "type": "string"
                            },
                            {
                                "name": "weekday_en",
                                "type": "string"
                            },
                            {
                                "name": "weekday_short",
                                "type": "string"
                            },
                            {
                                "name": "is_weekend",
                                "type": "boolean"
                            },
                            {
                                "name": "iso_week",
                                "type": "smallint"
                            },
                            {
                                "name": "month",
                                "type": "smallint"
                            },
                            {
                                "name": "month_cn",
                                "type": "string"
                            },
                            {
                                "name": "month_en_full",
                                "type": "string"
                            },
                            {
                                "name": "month_en_short",
                                "type": "string"
                            },
                            {
                                "name": "quarter",
                                "type": "string"
                            },
                            {
                                "name": "year",
                                "type": "smallint"
                            }
                        ],
                        "fieldDelimiter": "\t",
                        "maxFileSize": 134217728,
                        "encoding": "UTF-8"
                    }
                }
            }
        ],
        "setting": {
            "speed": {
                "channel": 5
            },
            "errorLimit": {
                "record": 0,
                "percentage": 0.02
            },
            "retry": {
                "limit": 3,
                "interval": 5000
            }
        }
    }
}