Datalore 2025.6 Help

Connect to an S3 bucket

The article explains how to create an S3 bucket connection.

Step 1. Create and configure a connection

  1. Open the New cloud storage connection dialog.

    1. On the Home page, select the workspace to which you want to add a cloud storage connection.

    2. In the expanded list of workspace resources, select Document Data and switch to the Cloud storage tab.

    3. Click Plus icon New cloud storage connection at the top right.

      Cloud storage tab
    1. Open the Attached data tool from the left-hand sidebar.

    2. Switch to the Cloud storage tab. You will see the list of all cloud storage connections available from the respective workspace.

    3. At the bottom of the tab, click New cloud storage.

      Creating cloud storage connection from notebook
  2. In the New cloud storage connection dialog, select Amazon S3.

  3. In the New Amazon S3 cloud storage connection dialog, fill in the following fields:

    • Display name: to specify the name for this data source in your system

    • AWS access key and AWS secret access key: to access your AWS account (details here)

    • Region: to specify your AWS region

    • Amazon Bucket name: to specify the name of the bucket you want to mount

    • Custom options: to specify additional parameters. See the example below

    • Custom endpoint URL: to specify the website of the bucket you want to mount

    New Amazon S3 cloud storage connection dialog
  4. (Optional) Click the Test connection button to make sure the provided parameters are correct.

  5. Click the Create and close button to finish the procedure.

Step 2. Configure optional parameters for S3 data sources

Use the Custom_options field for optional parameters when creating an Amazon S3 data source. Below are two examples of how it can be used.

  • To enable SSE-C for S3 data sources, specify the following in the Custom_options: In the Custom_options field, specify the following:

    use_sse=c:/path/to/keys/file

    where:

    /path/to/keys/file is the file that contain keys. Make sure permissions are 600.

  • (For Datalore On-Premises only) To provide access based on a role associated with that of an EC2 instance profile, add public_bucket=0,iam_role into the Custom_options field.

Step 3. Attach the connection to a notebook

  1. Open the notebook you want to attach the connection to.

  2. If the notebook is not running, start it by clicking Run | Start machine.

  3. In the sidebar, select Attached data and switch to the Cloud storage tab.

  4. Enable the toggle in your S3 connection.

  1. If the notebook is not running, start it by clicking Run | Start machine.

  2. In the sidebar, select Attached data and switch to the Cloud storage tab.

  3. Enable the toggle in your S3 connection.

    Step 4. Use the connection in the notebook

    After the connection is mounted, you can access files in the S3 bucket from your notebook code:

    import pandas as pd df = pd.read_csv("/data/s3/my_file.csv") df
    %use dataframe val df = DataFrame.readCsv("/data/s3/my_file.csv") df
    val lines = scala.io.Source.fromFile("/data/s3/my_file.txt").mkString
    df <- read.csv("/data/s3/my_file.csv") print(df)

    The connection’s mount path, such as /data/s3/, is shown on the connection card. To copy the path, click ∙∙∙ | Copy directory path.

      Next steps

      Keywords

      S3, bucket, cloud storage, cloud storage connection, attach data, data sources, Amazon S3

      24 November 2025