jumpman23
jumpman23

Reputation: 395

Use AWS Glue Python with NumPy and Pandas Python Packages

What is the easiest way to use packages such as NumPy and Pandas within the new ETL tool on AWS called Glue? I have a completed script within Python I would like to run in AWS Glue that utilizes NumPy and Pandas.

Upvotes: 17

Views: 46798

Answers (13)

SparkDataEng
SparkDataEng

Reputation: 89

Use Glue version 2 instead of version 3 Steps:

  1. Go to glue job and edit script with below code

code:

import sys
from awsglue.transforms import *
from awsglue.utils import getResolvedOptions
from pyspark.context import SparkContext
from awsglue.context import GlueContext
from awsglue.job import Job
import pandas as pd

args = getResolvedOptions(sys.argv, ['JOB_NAME'])

sc = SparkContext()
glueContext = GlueContext(sc)
spark = glueContext.spark_session
job = Job(glueContext)
job.init(args['JOB_NAME'], args)


excel_path= r"s3://input/employee.xlsx"
df_xl_op = pd.read_excel(excel_path,sheet_name = "Sheet1")
df=df_xl_op.applymap(str)
input_df = spark.createDataFrame(df)
input_df.printSchema()

job.commit()
  1. Save script

  2. Goto Action - Edit Job - Select Glue version2 and set key value under security configuration

    key : --additional-python-modules
    value : pandas==1.2.4,xlrd==1.2.0,numpy==1.20.1,fsspec==0.7.4

  3. Save and run the job

It will resolve your error and you will be able to read the excel file using pandas

Upvotes: 2

winnervc
winnervc

Reputation: 807

If you want to integrate python modules into your AWS GLUE ETL job, you can do. You can use whatever Python Module you want.

Because Glue is nothing but serverless with Python run environment. SO all you need is to package the modules that your script requires using pip install -t /path/to/your/directory. And then upload to your s3 bucket.

And while creating AWS Glue job, after pointing s3 scripts, temp location, if you go to advanced job parameters option, you will see python_libraries option there.

enter image description here

You can just point that to python module packages that you uploaded to s3.

Upvotes: 0

Suresh
Suresh

Reputation: 39501

AWS GLUE library/Dependency is little convoluted

there are basically three ways to add required packages

Approach 1

  1. via AAWS console UI/JOB definition, below are few screens to help
    Action --> Edit Job

    enter image description here

    then scroll all the way down and expand

    Security configuration, script libraries, and job parameters (optional)

    then add all your packages as .zip files to Python Library path (you need to add your .zip files to S3 then specify the path)

    one catch here is you need to make sure your zip file must contain init.py in the root folder

    enter image description here

and also, if your package depends on another package then it will be very difficult to add those packages.

Approach 2

programmatically installing your packages (Easy one)

here is the path where you can install the required libraries to

/home/spark/.local/lib/python3.7/site-packages/

**

/home/spark/.local/lib/python3.7/site-packages/

**

here is the example of installing the AWS package I have installed SAGE marker package here

import site
from importlib import reload 
from setuptools.command import easy_install
# install_path = site.getsitepackages()[0]
install_path = '/home/spark/.local/lib/python3.7/site-packages/'
easy_install.main( ["--install-dir", install_path, "https://files.pythonhosted.org/packages/60/c7/126ad8e7dfbffaf9a5384ca6123da85db6c7b4b4479440ce88c94d2bb23f/sagemaker-2.3.0.tar.gz"] )
reload(site)

Approach 3. (Suggested and clean)

under Security configuration, script libraries, and job parameters (optional) section to job parameters

add the required libraries with --additional-python-modules parameter you can specify as may packages as you need with comma separator enter image description here

happy to help

Upvotes: 3

Koo
Koo

Reputation: 1699

You can check latest python packages installed using this script as glue job

import logging
import pip
logger = logging.getLogger(__name__)
logger.setLevel(logging.INFO)

if __name__ == '__main__':
    logger.info(pip._internal.main(['list']))

As of 30-Jun-2020 Glue as has these python packages pre-installed. So numpy and pandas is covered.

awscli 1.16.242
boto3 1.9.203
botocore 1.12.232
certifi 2020.4.5.1
chardet 3.0.4
colorama 0.3.9
docutils 0.15.2
idna 2.8
jmespath 0.9.4
numpy 1.16.2
pandas 0.24.2
pip 20.0.2
pyasn1 0.4.8
PyGreSQL 5.0.6
python-dateutil 2.8.1
pytz 2019.3
PyYAML 5.2
requests 2.22.0
rsa 3.4.2
s3transfer 0.2.1
scikit-learn 0.20.3
scipy 1.2.1
setuptools 45.1.0
six 1.14.0
urllib3 1.25.8
virtualenv 16.7.9
wheel 0.34.2

You can install additional packages in glue-python if they are present in the requirements.txt used to build the attaching .whl. The whl file gets collected and installed before your script is kicked-off. I would also suggest you to look into Sagemaker Processing which is more easier for python based jobs. Unlike serveless instance for glue-python shell, you are not limited to 16gb limit there.

Upvotes: 15

victorx
victorx

Reputation: 3559

AWS Glue version 2.0 released on 2020 Aug now has pandas and numpy installed by default. See https://docs.aws.amazon.com/glue/latest/dg/reduced-start-times-spark-etl-jobs.html#reduced-start-times-new-features for detail.

Upvotes: 2

Kun
Kun

Reputation: 99

The picked answer is not longer true since 2019

awswrangler is what you need. It allows you to use pandas in glue and lambda

https://github.com/awslabs/aws-data-wrangler

Install using AWS Lambda Layer

https://aws-data-wrangler.readthedocs.io/en/latest/install.html#setting-up-lambda-layer

Example: Typical Pandas ETL

import pandas
import awswrangler as wr

df = pandas.read_...  # Read from anywhere

# Typical Pandas, Numpy or Pyarrow transformation HERE!

wr.pandas.to_parquet(  # Storing the data and metadata to Data Lake
    dataframe=df,
    database="database",
    path="s3://...",
    partition_cols=["col_name"],
)

Upvotes: 1

Sergey Nasonov
Sergey Nasonov

Reputation: 153

In order to install a specific version (for instance, for AWS Glue python job), navigate to the website with python packages, for example to the page of package "pg8000" https://pypi.org/project/pg8000/1.12.5/#files

Then select an appropriate version, copy the link to the file, and paste it into the snippet below:

import os
import site
from setuptools.command import easy_install
install_path = os.environ['GLUE_INSTALLATION']

easy_install.main( ["--install-dir", install_path, "https://files.pythonhosted.org/packages/83/03/10902758730d5cc705c0d1dd47072b6216edc652bc2e63a078b58c0b32e6/pg8000-1.12.5.tar.gz"] )
reload(site)

Upvotes: 0

Prabhakar Reddy
Prabhakar Reddy

Reputation: 5124

If you don't have pure python libraries and still want to use then you can use below script to use it in your Glue code:

import os
import site
from setuptools.command import easy_install
install_path = os.environ['GLUE_INSTALLATION']
easy_install.main( ["--install-dir", install_path, "<library-name>"] )
reload(site)


import <installed library>

Upvotes: 10

There is an update:

...You can now use Python shell jobs... ...Python shell jobs in AWS Glue support scripts that are compatible with Python 2.7 and come pre-loaded with libraries such as the Boto3, NumPy, SciPy, pandas, and others.

https://aws.amazon.com/about-aws/whats-new/2019/01/introducing-python-shell-jobs-in-aws-glue/

Upvotes: 6

BigData-Guru
BigData-Guru

Reputation: 1261

As of now, You can use Python extension modules and libraries with your AWS Glue ETL scripts as long as they are written in pure Python. C libraries such as pandas are not supported at the present time, nor are extensions written in other languages.

Upvotes: 1

Jasper_Li
Jasper_Li

Reputation: 244

I think the current answer is you cannot. According to AWS Glue Documentation:

Only pure Python libraries can be used. Libraries that rely on C extensions, such as the pandas Python Data Analysis Library, are not yet supported.

But even when I try to include a normal python written library in S3, the Glue job failed because of some HDFS permission problem. If you find a way to solve this, please let me know as well.

Upvotes: 11

letstry
letstry

Reputation: 83

when you click run job you have a button Job parameters (optional) that is collapsed by default , when we click on it we have the following options which we can use to save the libraries in s3 and this works for me :

Python library path

s3://bucket-name/folder-name/file-name

Dependent jars path

s3://bucket-name/folder-name/file-name

Referenced files path s3://bucket-name/folder-name/file-name

Upvotes: 2

MadCityDev
MadCityDev

Reputation: 316

If you go to edit a job (or when you create a new one) there is an optional section that is collapsed called "Script libraries and job parameters (optional)". In there, you can specify an S3 bucket for Python libraries (as well as other things). I haven't tried it out myself for that part yet, but I think that's what you are looking for.

Upvotes: 1

Related Questions