• Strong experience in migrating other databases to
Snowflake.
• Work with domain experts, engineers, and other
data scientists to develop,implement, and improve
upon existing systems.
• Experience in analyzing data using HiveSQL.
• Participate in design meetings for creation of the Data
Model and provide guidance on best data
architecture practices.
• Experience with Snowflake Multi - Cluster Warehouses.
• Experience in building Snowpipe.
• Experience in using Snowflake Clone and Time Travel.
• Experience in various methodologies like Waterfall and
Agile.
• Extensive experience in developing complex stored
Procedures/BTEQ Queries.
• In-depth understanding of Data Warehouse/ODS, ETL
concept and modeling structure principles
• Build the Logical and Physical data model for snowflake
as per the changes required
• Define roles, privileges required to access different
database objects.
• In-depth knowledge of Snowflake Database, Schema
and Table structures.
• Define virtual warehouse sizing for Snowflake for
different type of workloads.
• Worked with cloud architect to set up the environment
• Coding for Stored Procedures/ Triggers.
• Designs batch cycle procedures on major projects using
scripting and Control
• Develop SQL queries SnowSQL
• Develop transformation logic using snowpipeline.
• Optimize and fine tune queries
• Have good Knowledge in ETL and hands on experience
in ETL.
• Experience on Migrating SQL database to Azure data
Lake, Azure data lake Analytics, Azure SQL Database,
Data Bricks and Azure SQL Data warehouse and
Controlling and granting database access and
Migrating On premise databases to Azure Data lake
store using Azure Data factory.
• Analyze, design and build Modern data solutions using
Azure PaaS service to support visualization of data.
• Understand current Production state of application and
determine the impact of new implementation on
existing business processes.
• Extract Transform and Load data from Sources Systems
to Azure Data Storage services using a combination of
Azure Data Factory, T-SQL, Spark SQL and U-SQL
Azure Data Lake Analytics . Data Ingestion to one or
more Azure Services - (Azure Data Lake, Azure
Storage, Azure SQL, Azure DW) and processing the
data in In Azure Databricks.
• Operationalize data ingestion, data transformation and
data visualization for enterprise use.
• SAS Metadata and ETL developer with extensive
knowledge of building & implementing metadata
repository & metadata security
• Expertise in SAS Data Integration (DI) studio, SAS
Management console and SAS BI suite
• Strong knowledge of installation, configuration and
troubleshooting of SAS Information Map Studio, SAS
OLAP Cube Studio, SAS Information Delivery Portal and
SAS Web Report Studio
• SAS ETL developer with expertise in design and
development of Extract, Transform and Load processes
for data integration projects to build data marts
16+ years of experience on SAS Intelligence Platform
and its components.
• Statistical data analyst with extensive experience in
business data mining, database marketing and
statistical modeling using SAS.
• More than four years of experience as a SAS
Programmer and expert in data mining, database
marketing, direct marketing, predictive modeling,
customer profiling, clustering and segmentation
modeling using SAS Enterprise Miner & SAS EG.
• Extensive knowledge of using Base SAS, SAS/Macros,
SAS/SQL, SAS/STAT & SAS EG.
• Advanced knowledge of statistical models such as
clustering, segmentation, predictive modeling, ANOVA,
regression analysis, decision tree.
• Azure Storage, Azure SQL, Azure DW) and processing the data in In Azure Databricks. Pipelines were created in Azure Data Factory utilizing Linked Services/Datasets/Pipeline/ to extract, transform, and load data from many sources such as Azure SQL, Blob storage, Azure SQL Data warehouse, write-back tool, and backwards. Used Azure ML to build, test and deploy predictive analytics solutions based on data. Developed Spark applications wif Azure Data Factory and Spark-SQL for data extraction, transformation, and aggregation from different file formats in order to analyze and transform the data in order to uncover insights into customer usage patterns. Analyzed the SQL scripts and designed it by using PySpark SQL for faster performance. Applied technical knowledge to architect solutions that meet business, and IT needs, created roadmaps, and ensure long term technical viability of new deployments.
• Worked on SnowSQL and Snowpipe
• Created Snowpipe for continuous data load.
• Used COPY to bulk load the data.
• Created internal and external stage and transformed data during load.
• Redesigned the Views in snowflake to increase the performance.
• Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to
Extract, Transform and load data from different sources like Azure SQL,
Blob storage, Azure SQL Data warehouse, write-back tool and backwards.
• Developed stored process to perform deltas.
• Loaded the tables from the DWH to Azure data lake using azure data
factory integration run time.
• Loaded the tables from the azure data lake to azure blob storage for
pushing them to snowflake
• Migrating data from Teradata to Snowflake
• Experience in developing Spark applications using Spark-SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats for Analyzing transforming the data to uncover insights into the customer usage patterns
• Extract Transform and Load data from sources Systems to Azure Data Storage services using a combination of Azure Data factory, T-SQL
• Spark SQL and Azure Data Lake Analytics
• Data ingestion to one or more Azure services (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and processing the data in Azure Databricks
• Develop Spark applications using pyspark and spark SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing and transforming the data uncover insight into the customer usage patterns
• Extract, Parse cleaning and ingest data
• Responsible for estimating the cluster size, monitoring, and troubleshooting of the Spark databricks cluster
• Ability to apply the spark DataFrame API to complete Data manipulation within spark session
• Good understanding of Spark Architecture including spark core, spark SQL, DataFrame, Spark streaming, Driver Node, Worker Node, Stages, Executors and Tasks, Deployment modes, the Execution hierarchy, fault tolerance, and collection
• Experience in data ingestion and processing pipelines using spark and python
• Design ETL processes SAS Data Integration Studio to populate data from
various hospital information systems (Invitro , Vista and others) and the modelling of clinical workflow
• Build Jobs in DI Studio to support Ad hoc requests
• ETL Performance tuning, by using analytical functions and executing SQL within Teradata
• Build Stored Process in SAS EG with static and dynamic prompts for product-specific analyses
• SAS Stored process to Job execution services on SAS Viya
• Modifying existing codes for improvement in the performance
• Build visualizations and reports using the SAS Visual Analytics
• Build automated CAS User Transform to upload In Memory tables using
Delta's
• Migrate DI jobs from DI to SAS Viya ( import metadata , jobs create job service to replace existing stored process)
• Build & Deploy Flows and sub flows on LSF
• Environment: SAS 9.4, SAS Enterprise Guide 5.1, 7.1 SAS Data Integration Studio 4.9, SAS Information Map Studio 4.4, SAS Web Report Studio 4.4, Teradata 14.10, SAS The Cloud Analytic Services (CAS) Viya 3.4, Python 3.5.7, LSF(Platform Scheduler),ADF (Azure Data Factory) Databricks , Pyspark,Snowflake,Sql Sever