You are viewing a preview of this job. Log in or register to view more details about this job.

Data Engineer

Egen is a data engineering and cloud modernization firm helping industry-leading companies achieve digital breakthroughs and deliver for the future, today. We are catalysts for change who create digital breakthroughs at warp speed. Our team of cloud and data engineering experts are trusted by top clients in pursuit of the extraordinary. An Inc. 5000 Fastest Growing Company 7 times, and recently recognized on the Crain’s Chicago Business Fast 50 list, Egen has also been recognized as a great place to work 3 times.

Our Data Platform Engineering teams build scalable data pipelines using Python and AWS, GCP, or Azure. The pipelines we build typically integrate with technologies such as Kafka, Storm, and Elasticsearch. We are working on a continuous deployment pipeline that leverages rapid on-demand releases. Our developers work in an agile process to efficiently deliver high value applications and product packages.

As a Data Platform Engineer, you will architect and implement cloud-native data pipelines and infrastructure to enable analytics and machine learning on rich datasets.



Required Experience:

  • Minimum of Bachelor’s Degree or its equivalent in Computer Science, Computer Information Systems, Information Technology and Management, Electrical Engineering or a related field.
  • You know what it takes to build and run resilient data pipelines in production and have experience implementing ETL/ELT to load a multi-terabyte enterprise data warehouse.
  • You have implemented analytics applications using multiple database technologies, such as relational, multidimensional (OLAP), key-value, document, or graph.
  • You value the importance of defining data contracts, and have experience writing specifications including REST APIs.
  • You write code to transform data between data models and formats, preferably in Python (Spark or PySpark is a bonus).
  • You've worked in agile environments and are comfortable iterating quickly.

Nice to have's (but not required):

  • Experience moving trained machine learning models into production data pipelines.
  • Expert knowledge of relational database modeling concepts, SQL skills, proficiency in query performance tuning, and desire to share knowledge with others.
  • Experience building cloud-native applications and supporting technologies / patterns / practices including: AWS/GCP/Azure, Docker, CI/CD, DevOps, and microservices.