Big Data Engineer with Heavy SQL and ETL || Hybrid - mostly remote and occasionally onsite in Cleveland, OH || Client- State of Ohio - LOT at Cleveland, Ohio, USA |
Email: [email protected] |
http://bit.ly/4ey8w48 https://jobs.nvoids.com/job_details.jsp?id=659244&uid= From: Mohit Joshi, Exarca Inc [email protected] Reply to: [email protected] Role: Big Data Engineer with Heavy SQL and ETL Client: State of Ohio (LOT) Location : Hybrid (mostly remote and occasionally onsite in Cleveland, OH) Visa : USC, GC, GC-EAD, H4-EAD with LinkediN Contract Length: 1+ years Interview: Teams/Skype Job Description Title: Data Engineer The posting is hybrid/remote- occasional check-in at a lottery office located in the State of Ohio may be necessary. Skills Skill Required / Desired Amount of Experience Bachelor's or Masters degree in computer science, software or computer engineering, information systems or similar academic background. Required Strong knowledge of SQL to aid in data visualization Required 6 years Working knowledge of Hadoop tools, such as Spark, Impala, Hue, and Kafka, to create, query, and manage ETL data flows Required 6 years Working knowledge of data connectors for embedding data visualizations from Tableau Server or PowerBI into SharePoint Required Knowledge of Tableau Server and/or Microsoft PowerBI Required Python experience is a bonus Desired Working practical experience in developing and deploying ETL pipelines in an enterprise big data environment. Required Practical understanding and experience with data cleaning/cleansing. Required Conceptual understanding of networking schematics and data flows for ETL purposes Required Ability to conceptualize development of data mart environments using batch data for agency business analyst usage. Required Some experience with newer data analysis tools, such as Tensorflow, is helpful. Desired Responsibilities 85% Work with resources to provision necessary access to data sources for ingestion to agency Hadoop platform. Utilizing service accounts, develop ETL pipelines through Spark and/or StreamSets to bring data from disparate source systems into agencys Hadoop platform iteration. Validate data sets in Hadoop to ensure ingestion loads are as close to 1:1 as possible. Commit table transformations within Hadoop platform where needed to allow agency analyst to develop requested visuals. Ensure data refresh schedule is consistent with departmental need. Validate and test connections between dataset and cloud analytics software. Create a Data Mart DB for rolled-up and simplified tables for agency analyst consumption. One set of tables may be created for Sales staff, another may be created for Executive staff, etc. Further direction to be given once this stage is reached Assist with the automation of several existing reports and visualizations within Tableau Server Document work completed and publish FAQ on agency wiki. Work with department head to conduct agency training on data usage Validate and ingest streaming agency data, when available. Assist in transforming the data, with the data analyst, to create more on-demand, near-live visualizations of agency activity. Validate and test connections between dataset and cloud analytics software 10% Work with enterprise and corporate resources to determine how outside data and software (to mean outside of Microsoft ecosystem) can be read within Microsoft SharePoint server Deploy data connectors where necessary to connect IOP datasets to Microsoft ecosystem. Deploy data connectors, if necessary, to connect Tableau Server environment to Microsoft ecosystem 5% Document work completed and publish FAQ on agency wiki. Conduct training, as directed, throughout. Mohit Joshi Exarca Inc [email protected] Keywords: database green card Ohio http://bit.ly/4ey8w48 https://jobs.nvoids.com/job_details.jsp?id=659244&uid= |
[email protected] View All |
06:09 AM 20-Sep-23 |