Description:
Data Engineer - Level 3
Location/ Work Structure: Cincinnati, OH
Who we are:
Vernovis is a Total Talent Solutions company that specializes in Technology, Cybersecurity, Finance & Accounting functions. At Vernovis, we help these professionals achieve their career goals, matching them with innovative projects and dynamic direct hire opportunities in Ohio and across the Midwest.
Client Overview:
Vernovis partnering with a local Fortune 500 company to help identify talented Data Professionals for their growing Data Team. This is a great opportunity to join a well-established local company and help them transform their business.
What You'll Do:
- Working on data engineering pipelines, data modelling, analytics and supporting complex issues.
- Identify data acquisition strategies, design, and develop programs to acquire data into Data Lake/Data Warehouse
- Work with data team to efficiently use Google Cloud platform to analyze data, build data models and generate reports/visualizations
- Script Procedures to stage large amounts of input and output data in Data Lake; participate in software development life cycle (SDLC) tasks like status reporting, design reviews and approvals, detailed functional and technical documentation, migration of artifacts to production.
What Experience You'll Have:
- Bachelor’s degree in Statistics, Mathematics, Data Science, Engineering or related quantitative field
- Minimum 5 years of experience in Information Technology, including experience with SQL, Python, Big Query, Cloud Composer, Cloud Pub/Sub
- Minimum 5 years of experience in building and operationalizing large-scale enterprise data solutions and applications, with the ability to build production data pipelines from data ingestion to consumption within a hybrid big data architecture.
- Minimum 4 years of experience working with CI/CD pipelines using Code repositories like Gitlab, GitHub, etc. and deployment tools like Cloud Build or Gitlab Runners.
- Minimum 3 years of combined experience in PySpark, Cloud Dataproc, Cloud Dataflow, Terraform, Hadoop, Hive, Apache Spark, Cloud Spanner, Cloud SQL and Data Fusion.
- Minimum 1 year of experience in Reporting, Data Analytics and building data pipelines
- Experience in any cloud platforms such as GCP, Azure, AWS, with preference to GCP.
- Advanced understanding of technical environment, including technical area development and architecture of SAP.
What Experience is Nice to Have:
- Master’s degree in Statistics, Mathematics, Data Science, Engineering or related quantitative field with 3 years of experience in Information Technology, including exposure to:
- SQL, Python, Big Query, Cloud Composer, Cloud Pub/Sub
- Reporting and data analytics, such as building and operationalizing large scale enterprise data solutions and applications, building production data pipelines from data ingestion to consumption within a hybrid big data architecture
- Working with CI/CD pipelines using Code repositories like Gitlab, GitHub and deployment tools like Cloud Build or Gitlab Runners
- Experience in any cloud platforms such as GCP, Azure, AWS, etc. with preference to GCP
- Understanding of technical environment, including technical area development and architecture of SAP
- Experience in medallion architecture with Google Cortex framework
- Experience with containerization technologies like Docker or Kubernetes
- Experience in PySpark, Cloud Dataproc, Cloud Dataflow, Terraform, Hadoop, Hive, Apache Spark, Cloud Spanner, Cloud SQL and Data Fusion