Home

Opening for Data Engineer- Remote at Remote, Remote, USA
Email: [email protected]
http://bit.ly/4ey8w48
https://jobs.nvoids.com/job_details.jsp?id=445460&uid=

From:

meenakshi bisht,

kpg99

[email protected]

Reply to:   [email protected]

Title- Data Engineer

Location-Remote

Duration-6+months

USC/GC ONLY

Please note looking for Databricks Certified Data Engineer if not 3 years working with Databricks is a must

Note:

Non certified in Databricks:  If the user has good/hands on 2 to 4 years experience, we should consider.

As a Principal Data Engineer  you will contribute directly to our mission to empower clinicians and health systems with digital solutions to make a difference in patients and clinicians' lives. It's ambitious workand you're not on your own.

You will play a key role in design, development and support of cloud based clinical data platform and data applications within the Enterprise Connectivity group of Healthcare Systems and Technologies business unit. The data platform will support data ingestion from various on-premise and cloud-based data sources and data processing by leveraging various big data tools. The data platform will also support various BI and ML applications to develop actionable insights that improve clinical outcomes, benefit customer experience.

You will focus on designing, developing, expanding, and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross functional teams. You will support our software developers, database architects, data analysts and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects.

Your team

You will be a part of the Enterprise Connectivity R&D team, which supports both new development and sustainment of existing connectivity and other digital solutions. You will also direclty work with teams from other business units to onboard and develop data applications to support their products on the data platform.  You will work with a cross functional team that consists of other engineering disciplines, operations, and quality assurance.  We encourage partnership, collaboration and prioritize building relationships with each other. Our teams support innovation and out-of-the-box thinking balanced with shared expertise and accountability.

What you will be doing

Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Azure Data Factory, Azure Databricks and other Azure big data technologies.

Build data governance framework including data cataloging, access management, data auditing and data lineage supporting security and privacy compliance requirements for the data platform containing Protected Health Information (PHI).

Setup data processing workflow leveraging Azure Databricks, Unity Catalog, Auto-Loader, in Apache Spark notebooks using python or SQL languages.

Monitor and optimize data processing workloads to reduce processing time and overall compute cost.

Design and develop scalable and cost-effective data storage architecture for Medallion data architecture.

Build analytics framework that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency, and other key business performance metrics.

Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.

Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.

Work with data and analytics experts to strive for greater functionality in our data systems.

What you will bring

Bachelors Degree (Software Engineering or comparable technical subject area)

Databricks Certified Data Engineer Associate V3 certification is highly desired.

5+ years of development experience in building and optimizing big data data pipelines, architectures, and data sets. with enterprise grade cloud data platform.

Build processes supporting data transformation, data structures, metadata, dependency, and workload management. A successful history of manipulating, processing, and extracting value from large, disconnected datasets.

Working knowledge of message queuing, stream processing, and highly scalable big data data stores.

Hands on experience in performance tuning and optimizing code running in Databricks environment.

Proficient in programming languages like Python, SQL, Scala.

Good understanding of SQL, T-SQL and/or PL/SQL.

Experience with big data tools: Apache Spark, Kafka, etc.

Experience with relational SQL and NoSQL databases, including Postgres/SQL server and MongoDB/Cosmo.

Strong interpersonal skills with the ability to lead and work collaboratively with others multi-functionally.

Ability to plan, estimate and accurately track deliverables.

Regards,

Meenakshi Bisht

Technical Recruiter

D: 609-357-5502

E: [email protected]

Note:: We are going through prime vendor.

Keywords: machine learning business intelligence rlang information technology green card procedural language
http://bit.ly/4ey8w48
https://jobs.nvoids.com/job_details.jsp?id=445460&uid=
[email protected]
View All
09:51 PM 25-Jul-23


To remove this job post send "job_kill 445460" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]


Time Taken: 1

Location: ,