* Abort sync if one of the part fails to copy to temp table
* Check for record size when copying data from s3 to redshift
* Handle big record in RedshiftInsertDestination too
The Redshift Copy strategy currently has it's part size set to 10 MB. Since S3 allows a file to be broken into max of 10k parts, this results in a 100GB table limit. A user is trying to sync a table of 115GB and running into this issue.
This makes the part size configurable so users can increase this size if needed.
Release all connectors affected by namespace change. Includes all JDBC sources and destinations.
Also add documentation for normalisation. Prerequisite to actually releasing 0.21.0-alpha.