Integrating with AWS Lambda & DynamoDB
  • 01 Feb 2023
  • 9 Minutos para leer
  • Oscuro
    Ligero
  • PDF

Integrating with AWS Lambda & DynamoDB

  • Oscuro
    Ligero
  • PDF

The content is currently unavailable in Spanish. You are viewing the default English version.
Resumen del artículo

Note
  • While these articles are meant to serve as helpful guides for our customers, 1:1 support for integrations with 3rd party tools is limited in support interactions
  • If you are experiencing setup issues or need help troubleshooting this integration, we recommend you search threads in the Community first
  • If your question doesn't appear to be asked in the Community, then post a new message on the Development & APIs board (Requires signing into the community)

Airtable's developer platform gives you plenty of options to create, read, update and delete records in a given base. Whether you choose the scripting extension or develop a custom extension of your own, Airtable's developer platform can help you send data to other services, where you can centralize your data, and make it available to other teams throughout your organization.

Overview

This guide covers how to send information from Airtable to AWS DynamoDB. By following the steps in this guide, you will create:

  1. An AWS Dynamo Database where you can send Airtable data
  2. An AWS Gateway REST API to access your database
  3. An AWS Lambda Function to process inputs from Airtable
  4. An Airtable base from this example
  5. Two Airtable Scripting extensions that allow you to create, update, and delete information in your AWS Dynamo Database

Using AWS with Airtable

For many teams, AWS is their preferred development environment and is the backbone of many of their internal and customer-facing systems. By integrating Airtable with your AWS environment, you can make your Airtable data available to these other services.

NOTE

This guide focuses on pushing data into DynamoDB via an AWS Lambda. By changing the Lambda code, you could also use this setup to move data into other AWS services, such as RDS, Redshift or S3.


What you will need


Step-by-Step Instructions

To complete this integration we'll need to do some setup in both Airtable and AWS. We'll first need to create the base where we want our data to live. Then we'll create the DynamoDB instance where our data will live. After that, we can build the AWS Lambda and API Gateway environments which will then host the bulk of our code. Finally, we can configure our Airtable base to make API requests to this API to push data and actions into Dynamo.

1. Create an Airtable Base

Create your Airtable base and configure your tables and fields to reflect the information you want to capture, we recommend starting with this example base.

4403576621719ScreenShot2021-07-01at73951PM.png

2. Create a Dynamo Database
  1. Login to your AWS account and navigate to DynamoDB
  2. Select "Create table", then name your table and your primary key .
  3. Press "Create"
Note
Your primary key will define what makes a given record unique. You should use the same value here as you do in your Airtable base to define unique records. We generally call our primary key “record_id”

4403552071319ScreenShot2021-07-01at112236AM.png

3. Create a Lambda and REST API to handle requests

In this step, we'll create a Lambda Function and build a REST API using AWS API Gateway so that our Lambda Function can send information to our DynamoDB. For a more in-depth explanation into how REST APIs work, and why we're building components like Resources and Methods, watch this quick video.

Create a Lambda Function

  1. Within your AWS account, navigate to Lambda
  2. Select "Create Function"
  3. Select "Author from scratch," add your basic information, and select "Create function". We'll revisit this later. 4403552187927ScreenRecording2021-07-01at113218AM.gif

Create a REST API via AWS API Gateway

  1. Within your AWS account, navigate to API Gateway
  2. Under "Choose an API type" find REST API and select "Build"
  3. Select "New API" and give your API a name

be782e6d-0638-44fe-9a55-1a61337c6299.gif

Create Resources

Now we can create the resources for our API. For this project, we'll create two resources.

  1. In the tool bar, click "Actions" to open the dropdown menu. In the menu, click "Create Resource"
  2. Create your first resource with the following details:

Resource Name: table

Resource Path: table_name
3. Within /{table_name} we want a resource for ID so that a user can update/delete an existing item by making a request to /{table_name}/{id}. Within /{table_name}, create your second resource with the following details:

Resource Name: id

Resource Path: {id}

Create Methods

Now that we've added our resources, we can create methods under each resource. Our Lambda code will handle most of the routing logic based on HTTP method. For each resource complete the following steps.

  1. Select the resource name in the Resources sidebar
  2. In the "Actions" menu select "Create Method"
  3. Select ANY and click the checkmark to confirm
  4. Select "Lambda Function" as your integration type
  5. Select "Use Lambda Proxy Integration"
  6. Name your Lambda Function
  7. Select Save

4403552811415Image2021-07-01at121336PM.jpg

Deploy your API Gateway

  1. After you've created methods for both resources, select "Deploy API" from the Actions menu. On first set up, AWS will ask you to set up a Deployment stage. We've called ours “default”

4403566511639ScreenShot2021-07-01at44359PM.png

Note
  • You must re-deploy your API Gateway any time you make changes
  • Keep a record of the Invoke URL as that’s what we actually use to make API requests
  1. After you've deployed your API, you'll be taken to the Stage Editor where you can find your Invoke URL. This will be our endpoint for our API calls. Test your API by making a request via cURL or Postman.

4403574847255ScreenShot2021-07-01at44415PM.png

4. Add Lambda to an Identity Access and Management (IAM) Role

Before the Lambda Function can write data to your Dynamo DB, it will need permission. This can be done in the Identity Access and Management Service (IAM) by assigning the Lambda to a role that has the ability to issue CRUD operations to your Dynamo DB instance.

  1. Within your AWS account, navigate to Identity Access and Management (IAM)
  2. In the sidebar, select "Roles" and select the "Create New Roles" button

4403566678295ScreenShot2021-07-01at50055PM.png
3. Select AWS service and Lambda

4403575036567ScreenShot2021-07-01at50622PM.png
4. Search for AmazonDynamoDBFullAccess and select the policy.

4403575049367ScreenShot2021-07-01at51538PM.png
5. Walk through the rest of the creation process and apply additional data as necessary.
6. Go back to the Lambda Function you created and select Permissions. Next to Execution role click “Edit.”

4403566976663ScreenShot2021-07-01at52005PM.png
7. In the settings panel, select the policy you just created, and press save.

4403566976919ScreenShot2021-07-01at52136PM.png

5. Write your Lambda Code

We can now go back to our Lambda and update its code to actually manage these incoming operations. We will configure:

  1. Create new record
  2. Update existing record
  3. Delete existing record
Note
We've provided some sample code [in this GitHub repository.](https://github.com/Airtable-Labs/airtable_lambda_dynamodb) Feel free to copy and configure it to your needs. You can download the repository as a ZIP and then upload the ZIP directly into Lambda.
  • Create will use the /{table_name} endpoint and take a JSON body payload.
  • Update and Delete will use the /{table_name}/{id} endpoint to perform an action on a specific record.

Our example code also assumes you have the following environment variables configured which you can set in the Environment Variables section underneath the Lambda code editor.

CORS_ORIGIN = example.com
- The domains that you will allow cross-origin requests from
- If you know the subset of domains, you can put them here to decrease exposure

IS_CORS = true
- Can set to false if you do not want to allow cross-origin requests

PRIMARY_KEY = record_id
- This should be the DynamoDB key name that you set when creating or DynamoDB instance.

TIP

Allowing cross-origin requests is necessary for this solution to work, but you should familiarize yourself with the risks before doing so.

6. Add Scripting extension to your Airtable Base

Airtable’s scripting extension allows you to write JavaScript or TypeScript code and deploy it from your base. This enables you to further optimize your workflows by including custom logic in your base.

The scripting extension supports the ability to make web-requests to other services via a fetch request. Allowing us to make API calls out to our new AWS API Gateway and Lambda application, which will create and delete data within our DynamoDB database.

4403568805399ScreenShot2021-07-01at72138PM.png

Airtable also supports a button field , whichcan be configured to take a number of different actions after the user clicks on it. One of those actions is the ability to pass a record’s details to a scripting extension and then execute the script with those details.

For our example use case, we want users to consciously make the decision to push updates into DynamoDB vis a button in the Airtable base. In our example, Airtable users plan, draft and collaborate on content in a base. Once it’s ready to publish into our DyanmoDB instance, the user can push a button to send that record into DynamoDB. If the user needs to pull the information from DynamoDB because it’s no longer relevant, users can trigger a deletion by pushing another button in the base.

4403568820247ScreenShot2021-07-01at72246PM.png

To configure this workflow you’ll first need to create your scripts and then create the button fields which can be used to trigger those scripts.

Create / Update Script

Create a new scripting extension in your base and use the following code:

const base_url =  '{{YOUR AWS API GATEWAY URL}}' ; 
const dynamo_table_name =  '{{YOUR DYNAMODB TABLE NAME}}' ; 
const primary_dynamo_field =  '{{YOUR DYNAMODB PRIMARY FIELD/KEY NAME}}' ; 


// build the URL 

const lambda_url = base_url + dynamo_table_name; 


// get our different tables   

const table = base.getTable( 'Table 1' ); 
const foreign_table = base.getTable( 'Table 2' ); 


// get the record from the button action,   
// or allow the user to run the script and pick a specific record 

const record =  await input.recordAsync( "Select record to update" , table);

/** 

* For a given record, convert it into a JSON object 

* This function iterates over all fields in a record and converts it into a format that we can then send to Dynamo 

* Most of the logic is managing linked record fields.     

* We recursively call the function on any linked records, so that we get all of the linked records' information as well 
*/ 

async function recordToJson(table, record, linked_table_id) { 
 var p = {}; 
 for ( var f  of table.fields) { 
 if (f. type ===  'multipleRecordLinks' && f.options.linkedTableId !== linked_table_id) { 
 var foreign   = base.getTable(f.options.linkedTableId); 
 var res =  await foreign.selectRecordsAsync(); 
 var foreign_recs = record.getCellValue(f.id); 

foreign_recs = foreign_recs ===  null ? [] : foreign_recs 
p[f.name] = []; 

 for ( var r  of foreign_recs){ 

p[f.name].push( await recordToJson(foreign, res.getRecord(r.id), table.id)); 
} 

p[f.name] =  JSON .stringify(p[f.name]); 
}  
else if (f. type !==  'button' && f. type !==  'multipleRecordLinks' ) { 
p[f.name] = record.getCellValueAsString(f);                 
} 

} 

 return p; 
} 


// build the payload to send and then send to our Lambda 

var payload =  await recordToJson(table, record); 
var p =  await fetch( 

lambda_url,   

{ 
method: 'POST' , 
body:  JSON .stringify(payload), 
headers: { 
 'Content-Type' :  'application/json' 
} 

} 

); 

var resp =  await p.json(); 

output.markdown( 'Update sent!' ); 
output.inspect(resp); 


// then write the last updated time for this record back to the table 

output.markdown( 'Updating time sent' ); 

await table.updateRecordAsync( 
record.id,   
{ 

 'Last Pushed to Dynamo' : ( new Date ()).toISOString() 

} 

); 

output.markdown( '## Done!' ); 

Delete Script

Create another scripting block and use the following code:

const base_url =  '{{YOUR AWS API GATEWAY URL}}' ; 
const dynamo_table_name =  '{{YOUR DYNAMODB TABLE NAME}}' ; 
const primary_dynamo_field =  '{{YOUR DYNAMODB PRIMARY FIELD/KEY NAME}}' ; 
const table = base.getTable( 'Table 1' ); 
const record =  await input.recordAsync( "Select record to update" , table); 
const url =    ${base_url} / ${dynamo_table_name} / ${record.getCellValueAsString(primary_dynamo_field)}   ; 

var p =  await fetch( 
url,   

{ 
method: 'DELETE' , 
headers: { 
 'Content-Type' :  'application/json' 
} 
} 

); 

var resp =  await p.json(); 
output.markdown( 'Record deleted!' ); 
output.inspect(resp); 

await table.deleteRecordAsync(record.id); 

output.markdown( 'Deleted record from base' ); 

Create the Button Fields

To wrap this all up, we can now create two button fields:

  1. Create/Update Record
  2. Delete Record

When creating the field:

  1. Action : Run Script
  2. Dashboard : the dashboard you installed the scripting extension in
  3. Extension : the corresponding update/delete script

You can then trigger these scripts by clicking on the buttons and see the corresponding data in your AWS DynamoDB instance.



¿Te ha sido útil este artículo?