to_s3

DataFrame.to_s3(bucket: str, key: str, region: str, sep=',', batch_size=50000)[source]

Writes the underlying data into a newly created CSV file located in an S3 bucket.

NOTE THAT S3 IS NOT SUPPORTED ON WINDOWS.

Args:
bucket (str):

The bucket from which to read the files.

key (str):

The key in the S3 bucket in which you want to write the output. The ending “.csv” and an optional batch number will be added automatically.

region (str):

The region in which the bucket is located.

sep (str, optional):

The character used for separating fields.

batch_size(int, optional):

Maximum number of lines per file. Set to 0 to read the entire data frame into a single file.

Example:
>>> getml.engine.set_s3_access_key_id("YOUR-ACCESS-KEY-ID")
>>>
>>> getml.engine.set_s3_secret_access_key("YOUR-SECRET-ACCESS-KEY")
>>>
>>> your_df.to_s3(
        ...     bucket="your-bucket-name",
        ...     key="filename-on-s3",
        ...     region="us-east-2",
        ...     sep=';'
        ... )