Lokesh - Sr. Data Engineer |
[email protected] |
Location: Pittsburgh, Pennsylvania, USA |
Relocation: Any were in USA |
Visa: H1B |
Resume file: Lokeshm_DataEngineer_1744990390207.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
PROFESSIONAL SUMMARY
Data Engineer with 7+ Plus years of experience in Designing, Developing, Documentation, and integrating applications using Big Data platforms like Cloudera, Hortonworks, and Map Reduce, using cloud technologies Mainly AWS. Strong experience in Data Analyst, Data mining with large data sets of Structured and Unstructured data, Data Acquisition, Data Validation, Predictive modeling, Statistical modeling, Data modeling, Data Visualization, Web Crawling, Web Scraping. Adept in statistical programming languages like R and Python, SAS, Apache Spark, Matlab including Big Data technologies like HDFS, Map Reduce, Hive, HBase, Pig, Spark, Spark Streaming, Impala, Oozie, Flume, Catalog, Kafka, and Sqoop. Experience with all SDLC and Agile Development model stages right from requirement gathering to deployment and production support Experience in configuring the Zookeeper to coordinate the servers in clusters and to maintain data consistency Experience on Apache Oozie for scheduling and managing the Hadoop Jobs Experience with extracting Real-time feed using Kafka and Spark Streaming, converting it to RDD, processing data in the form of Data Frame, and saving the data as Parquet format in HDFS Experience in working with MapReduce programs using Apache Hadoop for working with Big Data. Extensively worked with Teradata utilities Fast export, and Multi Load to export and load data to/from different source systems including flat files Experience with visualizing the data using BI and services amp tools such as Power BI, Tableau and Matplotlib. Experience working on various file-formats including delimited text files, clickstream log files, Apache log files, Parquet files, Avro files, JSON files, XML files, and others. Experience in data mart life cycle development, performed ETL procedure to load data from various sources into Data marts and Data warehouses using Informatica Power Center. Expertise in NOSQL databases and its integration with Hadoop cluster to store and retrieve huge amount of data. Experience in writing SQL queries and optimizing the databases like MySQL, SQL Server, and Oracle. Hands-on experience using Airflow DAG for Managing and scheduling Jobs on a Hadoop cluster. Good understanding of data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Schema Modeling, Fact and Dimension tables. Strong knowledge of NoSQL Databases like MongoDB, HBase, Oracle, Cosmos DB, and Dynamo DB Experience in UNIX Shell Scripting and Deployment of Applications in Server. Involved in Agile methodologies, daily Scrum meetings, Sprint Planning, and sprint retrospective meetings Experience in creating Reports and Dashboards using Power BI on the requirement of the team Experienced in AWS Cloud Computing services, such as EC2, S3, Lambda, API, DynamoDB, EBS, VPC, ELB, Route53, Cloud Watch, Security Groups, Cloud Trail, IAM, Cloud Front, EMR, RDS and Glacier also worked on DNS, SSL and Firewalls. Keywords: business intelligence sthree database rlang information technology |