In the present day, we’re happy to announce a brand new AWS Glue connector for Google Cloud Storage that lets you transfer information bi-directionally between Google Cloud Storage and Amazon Easy Storage Service (Amazon S3).
We’ve seen that there’s a demand to design purposes that allow information to be transportable throughout cloud environments and provide the skill to derive insights from a number of information sources. One of many information sources now you can rapidly combine with is Google Cloud Storage, a managed service for storing each unstructured information and structured information. With this connector, you possibly can deliver the information from Google Cloud Storage to Amazon S3.
On this publish, we go over how the brand new connector works, introduce the connector’s capabilities, and offer you key steps to set it up. We offer you stipulations, share the way to subscribe to this connector in AWS Market, and describe the way to create and run AWS Glue for Apache Spark jobs with it.
AWS Glue is a serverless information integration service that makes it easy to find, put together, and mix information for analytics, machine studying, and utility improvement. AWS Glue natively integrates with numerous information shops resembling MySQL, PostgreSQL, MongoDB, and Apache Kafka, together with AWS information shops resembling Amazon S3, Amazon Redshift, Amazon Relational Database Service (Amazon RDS), Amazon DynamoDB, and Amazon S3. AWS Glue Market connectors help you uncover and combine further information sources, resembling software program as a service (SaaS) purposes and your customized information sources. With only a few clicks, you possibly can search and choose connectors from AWS Market and start your information preparation workflow in minutes.
How the connector works
This connector depends on the Spark DataSource API in AWS Glue and calls Hadoop’s FileSystem interface. The latter has carried out libraries for studying and writing numerous distributed or conventional storage. This connector additionally consists of the Google Cloud Storage Connector for Hadoop, which helps you to run Apache Hadoop or Apache Spark jobs immediately on information in Google Cloud Storage. AWS Glue hundreds the library from the Amazon Elastic Container Registry (Amazon ECR) repository throughout initialization (as a connector), reads the connection credentials utilizing AWS Secrets and techniques Supervisor, and reads information supply configurations from enter parameters. When AWS Glue has web entry, the Spark job in AWS Glue can learn from and write to Google Cloud Storage.
Answer overview
The next structure diagram exhibits how AWS Glue connects to Google Cloud Storage for information ingestion.
Within the following sections, we present you the way to create a brand new secret for Google Cloud Storage in Secrets and techniques Supervisor, subscribe to the AWS Glue connector, and transfer information from Google Cloud Storage to Amazon S3.
Stipulations
You want the next stipulations:
- An account in Google Cloud and your information path in Google Cloud Storage. Put together the GCP account keys file prematurely and add them to the S3 bucket. For directions, consult with Create a service account key.
- A Secrets and techniques Supervisor secret to retailer a Google Cloud secret.
- An AWS Identification and Entry Administration (IAM) function for the AWS Glue job with the next insurance policies:
- AWSGlueServiceRole, which permits the AWS Glue service function entry to associated companies.
- AmazonEC2ContainerRegistryReadOnly, which gives read-only entry to Amazon EC2 Container Registry repositories. This coverage is for utilizing AWS Market’s connector libraries.
- A Secrets and techniques Supervisor coverage, which gives learn entry to the key in Secrets and techniques Supervisor.
- An S3 bucket coverage for the S3 bucket that it’s essential load ETL (extract, rework, and cargo) information from Google Cloud Storage.
We assume that you’re already conversant in the important thing ideas of Secrets and techniques Supervisor, IAM, and AWS Glue. Concerning IAM, these roles needs to be granted the permissions wanted to speak with AWS companies and nothing extra, in line with the precept of least privilege.
Create a brand new secret for Google Cloud Storage in Secrets and techniques Supervisor
Full the next steps to create a secret in Secrets and techniques Supervisor to retailer the Google Cloud Storage credentials:
- On the Secrets and techniques Supervisor console, select Retailer a brand new secret.
- For Secret sort, choose Different sort of secret.
- Enter your key as
keyS3Uri
and the worth as your key file within the s3 bucket, for instance,s3://keys/project-gcs-connector **.json
. - Go away the remainder of the choices at their default.
- Select Subsequent.
- Present a reputation for the key, resembling
googlecloudstorage_credentials
. - Observe the remainder of the steps to retailer the key.
Subscribe to the AWS Glue connector for Google Cloud Storage
To subscribe to the connector, full the next steps:
- Navigate to the Google Cloud Storage Connector for AWS Glue on AWS Market.
- On the product web page for the connector, use the tabs to view details about the connector. In case you resolve to buy this connector, select Proceed to Subscribe.
- Evaluation the pricing phrases and the vendor’s Finish Consumer License Settlement, then select Settle for Phrases.
- Proceed to the following step by selecting Proceed to Configuration.
- On the Configure this software program web page, select the achievement choices and the model of the connector to make use of. We’ve got offered two choices for the Google Cloud Storage Connector, AWS Glue 3.0 and AWS Glue 4.0. On this instance, we deal with AWS Glue 4.0. After choosing Glue 3.0 or Glue 4.0, choose corresponding AWS Glue model whenever you configure the AWS Glue job.
- Select Proceed to Launch.
- On the Launch this software program web page, you possibly can assessment the Utilization Directions offered by AWS. While you’re able to proceed, select Activate the Glue connector in AWS Glue Studio.
The console will show the Create market connection web page in AWS Glue Studio.
Transfer information from Google Cloud Storage to Amazon S3
To maneuver your information to Amazon S3, you need to configure the customized connection after which arrange an AWS Glue job.
Create a customized connection in AWS Glue
An AWS Glue connection shops connection info for a selected information retailer, together with login credentials, URI strings, digital personal cloud (VPC) info, and extra. Full the next steps to create your connection:
- On the AWS Glue console, select Connectors within the navigation pane.
- Select Create connection.
- For Connector, select Google Cloud Storage Connector for AWS Glue.
- For Title, enter a reputation for the connection (for instance,
GCSConnection
). - Enter an non-obligatory description.
- For AWS secret, enter the key you created (
googlecloudstorage_credentials
). - Select Create connection and activate connector.
The connector and connection info is now seen on the Connectors web page.
Create an AWS Glue job and configure connection choices
Full the next steps:
- On the AWS Glue console, select Connectors within the navigation pane.
- Select the connection you created (
GCSConnection
). - Select Create job.
- On the Node properties tab within the node particulars pane, enter the next info:
- For Title, enter Google Cloud Storage Connector for AWS Glue. This title needs to be distinctive amongst all of the nodes for this job.
- For Node sort, select the Google Cloud Storage Connector.
- On the Information supply properties tab, present the next info:
- For Connection, select the connection you created (
GCSConnection
). - For Key, enter path, and for Worth, enter your Google Cloud Storage URI (for instance,
gs://bucket/covid-csv-data/
). - Enter one other key-value pair. For Key, enter
fileFormat
. For Worth, entercsv
, as a result of our pattern information is that this format.
- For Connection, select the connection you created (
- On the Job particulars tab, for IAM Function, select the IAM function talked about within the stipulations.
- For Glue model, select your AWS Glue model.
- Proceed to create your ETL job. For directions, consult with Creating ETL jobs with AWS Glue Studio.
- Select Run to run your job.
After the job succeeds, we will test the logs in Amazon CloudWatch.
The information is ingested into Amazon S3, as proven within the following screenshot.
We are actually capable of import information from Google Cloud Storage to Amazon S3.
Scaling issues
On this instance, we set the AWS Glue capability as 10 DPU (Information Processing Models). A DPU is a relative measure of processing energy that consists of 4 vCPUs of compute capability and 16 GB of reminiscence. To scale your AWS Glue job, you possibly can improve the variety of DPU, and likewise make the most of Auto Scaling. With Auto Scaling enabled, AWS Glue routinely provides and removes employees from the cluster relying on the workload. This removes the necessity so that you can experiment and resolve on the variety of employees to assign in your AWS Glue ETL jobs. In case you select the utmost variety of employees, AWS Glue will adapt the proper measurement of sources for the workload.
Clear up
To wash up your sources, full the next steps:
- Take away the AWS Glue job and secret in Secrets and techniques Supervisor with the next command:
- Cancel the Google Cloud Storage Connector for AWS Glue’s subscription:
- On the AWS Market console, go to the Handle subscriptions web page.
- Choose the subscription for the product that you simply need to cancel.
- On the Actions menu, select Cancel subscription.
- Learn the knowledge offered and choose the acknowledgement test field.
- Select Sure, cancel subscription.
- Delete the information within the S3 buckets.
Conclusion
On this publish, we confirmed the way to use AWS Glue and the brand new connector for ingesting information from Google Cloud Storage to Amazon S3. This connector gives entry to Google Cloud Storage, facilitating cloud ETL processes for operational reporting, backup and catastrophe restoration, information governance, and extra.
This connector allows your information to be transportable throughout Google Cloud Storage and Amazon S3. We welcome any suggestions or questions within the feedback part.
References
Concerning the authors
Qiushuang Feng is a Options Architect at AWS, answerable for Enterprise prospects’ technical structure design, consulting, and design optimization on AWS Cloud companies. Earlier than becoming a member of AWS, Qiushuang labored in IT firms resembling IBM and Oracle, and gathered wealthy sensible expertise in improvement and analytics.
Noritaka Sekiyama is a Principal Huge Information Architect on the AWS Glue staff. He’s captivated with architecting fast-growing information environments, diving deep into distributed large information software program like Apache Spark, constructing reusable software program artifacts for information lakes, and sharing information in AWS Huge Information weblog posts.
Greg Huang is a Senior Options Architect at AWS with experience in technical structure design and consulting for the China G1000 staff. He’s devoted to deploying and using enterprise-level purposes on AWS Cloud companies. He possesses almost 20 years of wealthy expertise in large-scale enterprise utility improvement and implementation, having labored within the cloud computing subject for a few years. He has in depth expertise in serving to numerous varieties of enterprises migrate to the cloud. Previous to becoming a member of AWS, he labored for well-known IT enterprises resembling Baidu and Oracle.
Maciej Torbus is a Principal Buyer Options Supervisor inside Strategic Accounts at Amazon Internet Providers. With in depth expertise in large-scale migrations, he focuses on serving to prospects transfer their purposes and programs to extremely dependable and scalable architectures in AWS. Exterior of labor, he enjoys crusing, touring, and restoring classic mechanical watches.