Skip to content

Instantly share code, notes, and snippets.

Show Gist options
  • Save AvinashDalvi89/f028811f3b5b443f23e65bf95944728e to your computer and use it in GitHub Desktop.
Save AvinashDalvi89/f028811f3b5b443f23e65bf95944728e to your computer and use it in GitHub Desktop.
This include angular build creation, upload to S3 and clearing cloudfront cache.
#!/usr/bin/env python
"""
File : deployment_angular_s3_cloudfront.py
Package :
Description :
Created by Avinash on 18/09/2019
"""
# git checkout production
# git fetch origin
# git merge origin/production
# ng build --prod --aot=false
# ng build --configuration=production
# take bakup file of
# inline.*
# main.*
# polyfills.*
# scripts.*
# styles.*
# vendor.*
# create folder "backup_current date" -- backup_18092019. - optional step for rollback deployment.
# cut this above files and paste into backup folder
# Copy files from "dist" folder all files and paste into root folder
from functools import wraps
from zipfile import is_zipfile
import os
import random
import string
import subprocess
import datetime
import boto3
from constants import *
from pygit2 import Repository
import mimetypes
class DeploymentService():
def __init__(self,request):
self.bucket_name = request.get('bucketName')
self.env = request.get('envName')
self.build_path = request.get('buildPath')
self.repo_path = request.get('repositoryPath')
self.connect_aws()
self.distribution_id = request.get('distributionId')
self.portal = request.get('portal')
self.build_command = request.get('buildCommand')
def get_file_name(self,input):
return input.split('/')[-1]
def get_folder_name(self,input):
return input.split('/')[-2]
def connect_aws(self):
self.aws_session = boto3.Session(profile_name=AWS_CONFIG.get(self.env))
self.s3_client = self.aws_session.client('s3')
self.s3_resource = self.aws_session.resource('s3')
def start_process(self):
folder_list, file_list = list_object = self.list_objects()
backup_folder_name = "backup_"+ datetime.datetime.now().strftime("%d%m%Y")
if backup_folder_name not in folder_list:
self.s3_client.put_object(Bucket=self.bucket_name, Key=(backup_folder_name + '/'))
for file in file_list:
file_header_name = file.split(".")
if file_header_name[0] in PROD_EXPECTED_LIST.get(self.portal) and file_header_name[0] not in EXCLUDE_LIST.get(self.portal):
copy_source = {'Bucket': self.bucket_name, 'Key': file}
self.s3_client.copy_object(CopySource=copy_source, Bucket=self.bucket_name, Key=backup_folder_name+"/"+file)
self.s3_client.delete_object(Bucket=self.bucket_name, Key=file)
self.create_build()
self.upload_build()
self.clear_cache_invalidation(self.distribution_id)
def list_objects(self):
list_object = self.s3_client.list_objects_v2(Bucket=self.bucket_name)
file_array = []
#print(list_object)
folder_array = []
if 'Contents' in list_object.keys():
for list in list_object['Contents']:
# print list
if "/" in list.get('Key'):
folder_name = self.get_folder_name(list.get('Key'))
folder_array.append(folder_name)
else:
file_name = self.get_file_name(list.get('Key'))
file_array.append(file_name)
# print(folder_array)
# print(file_array)
return folder_array,file_array
def create_build(self):
current_branch = Repository(self.repo_path).head.shorthand
pwd = self.repo_path
os.chdir(pwd)
os.system('git fetch origin')
if current_branch != BRANCH_CONFIG.get(self.env):
os.system('git checkout %s' % BRANCH_CONFIG.get(self.env))
os.system('git merge origin/%s' % BRANCH_CONFIG.get(self.env))
print("Build starting...")
build_pwd = self.build_path
os.chdir(build_pwd)
print("Running %s" % self.build_command)
os.system(self.build_command)
print("Build Completed...")
def upload_build(self):
try:
print("Uploading code to S3 is in progress...")
root_path = "%s/dist" % self.build_path
print(root_path)
image_dir = False
for root, dirs, files in os.walk(root_path):
if 'images' in dirs:
image_dir = True
for file in files:
s3_file = file
if image_dir:
s3_file = "assets/images/%s" % file
read_mime_types = mimetypes.guess_type(s3_file)
if read_mime_types[0]:
content_type = read_mime_types[0]
else:
content_type = 'binary/octet-stream'
self.s3_client.upload_file(os.path.join(root, file), self.bucket_name, s3_file,ExtraArgs={'ContentType':content_type})
print("Code uploaded successfully!")
except Exception as err:
print(err)
def clear_cache_invalidation(self,distribution_id):
unique_time_stamp = datetime.datetime.timestamp(datetime.datetime.now())
print(unique_time_stamp)
client = self.aws_session.client('cloudfront')
try:
response = client.create_invalidation(
DistributionId= distribution_id,
InvalidationBatch={
'Paths': {
'Quantity': 2,
'Items': [
'/*','/index.html'
]
},
'CallerReference': str(unique_time_stamp)
}
)
if response.get('ResponseMetadata').get('HTTPStatusCode') == 201 or response.get('HTTPStatusCode') == 200:
print("Invalidation Request Initiated...")
print('To check status go to invalidation tab.')
#print(response)
except Exception as err:
print(err)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment