In this simple guide, we'll run through how to import AWS S3 Data to Airtable without the need for any code.
AWS (Amazon Web Services) is a cloud computing platform by Amazon. AWS provides various scalable services, including computing power, storage, databases, networking and machine learning. By using AWS, businesses and individuals can easily access and utilize computing resources on-demand, paying only for what they use.
We'll be importing data from an AWS service called S3 or Amazon Simple Storage Service. S3 is a cloud-based storage service similar to Dropbox or iCloud.
We'll be using Data Fetcher to import AWS S3 Data to Airtable. Data Fetcher is a powerful Airtable extension that enables you to import many different types of data from APIs or websites into Airtable.
Add Data Fetcher to Airtable via the Airtable marketplace. Sign up for a Data Fetcher account by entering a password and clicking 'Sign up for free'. Alternatively, you can use your Google login to create a new account. If you already have a Data Fetcher account, use the 'Have an account?' text in the bottom left of the screen to log in.
Data Fetcher requests enable you to import data into your Airtable base. To begin, click 'Create your first request' on the home screen of the Data Fetcher extension.
For Application, select 'Custom' to set up the AWS Airtable integration.
For this demo, we'll use an example CSV file of UK government jobs data, which you can find here (or use your own CSV file.) This needs to be uploaded to your S3 bucket. You can follow these instructions if you are unsure of this process.
We will then use the AWS S3 API to import our CSV file into Airtable.
The URL to get the file using the API is: https://s3.amazonaws.com/jobs-bucket-data-fetcher/jobs.csv
`jobs-bucket-data-fetcher` is our S3 bucket name, so replace 'jobs-bucket-data-fetcher' with your S3 bucket name.
`jobs.csv` is our file name, so replace this with your file name. You might also need to add a folder name to the URL before the file name.
Enter your URL into the URL field in Data Fetcher.
Under Authorization, select 'AWS Signature V4' and enter your AWS Access key and Secret key. For Region, use the region that your S3 bucket is in 'us-east-1'.
You can find out how to access your AWS keys here. Alternatively, you can use a root user access key. However, it's better security practice to use IAM users with the minimum level of security.
This authorization method will also work with any other AWS API.
Give your request a name, such as 'Import AWS S3 Data' and click 'Save & Run'.
Data Fetcher will run the request to AWS, and the Response field mapping modal will open. This is where we can select which fields from the AWS file to import into Airtable, and we can choose how they map to our output table.
It is possible to select 'Filter all' to clear all the pre-selected fields and then use the Find field search bar to locate the fields you want to import.
Once you are happy, click 'Save & Run'.
Data Fetcher will create any new fields in the output table. It will then run the request to AWS and import the data from the CSV file into Airtable. You'll now be able to see the data in your output table.
Currently, you'd need to manually run the request each time you want to import any new data from your AWS S3 file. It is possible to automatically run the AWS Airtable integration on a chosen schedule with Data Fetcher's scheduling feature.
Scheduling is a paid feature, so in Data Fetcher, scroll to Schedule and click 'Upgrade'.
Choose the plan that suits you the best and enter your payment details.
In Data Fetcher, click 'I've done this'.
Under Schedule click '+ Authorize'.
A window will now open where you'll need to authorize the Airtable bases you want Data Fetcher access.
By selecting 'All current and future bases in all current and future workspaces' you'll also avoid having to authorize bases for future use.
Click 'Grant access'.
In Data Fetcher, you'll see Schedule this request is now toggled to on.
Select a schedule for the AWS Airtable integration to run. You can choose intervals of 'Minutes', 'Hours', 'Days' or 'Months', then click 'Save'.
If changes are made to the file you have stored in AWS S3 these will now all be imported into Airtable automatically on this schedule.
Dec 10, 2024
•Andy Cloke
•WebflowWhalesyncOct 7, 2024
•Zayyad Muhammad Sani
•Custom RequestsData FetcherSep 20, 2024
•Zayyad Muhammad Sani
•Custom Requests