Enroute is about being exceptional. We deliver IT services and solutions by tech-savvy problem solvers, constantly looking for innovative approaches to everyday problems. Enrouters have unique ideologies, principles, and incredible life stories. Everyone is welcome at Enroute.
We take pride in our culture. We want every Enrouter to enjoy working with us and become part of a great community of highly driven, responsible, respectful, and happy people. We offer outstanding benefits, compensation, flexible schedules, and policies that balance work and personal life. We strive to be involved and know our people to improve continuously.
We are looking for a Data Engineer who will vital in designing, enhancing, and maintaining the infrastructure and data pipelines for efficient data processing and analysis.
Requirements
- Python
- SQL
- Spark, PySpark, Pandas.
- AWS services (Glue, S3, Step Functions)
- Git
- CI/CD
- Bash
Responsibilties
- Data Pipeline Development: You will design, build, and refine data pipelines for structured and unstructured data types by leveraging technologies like Python, SQL, Spark, PySpark, Pandas, or similar technologies or frameworks to ensure the accuracy, quality, and consistency of data.
- Cloud Integration: Employ cloud services like AWS Glue, S3, and Step Functions or similar on other cloud providers like GCP and Azure to manage data workflows, efficiently extract, load, and transform data, and ensure seamless integration with existing cloud infrastructure.
- Data Modeling: Collaborate with data scientists and other stakeholders to understand data requirements and design appropriate data models. Implement data structures that support efficient querying and analysis while ensuring data integrity and security.
- ETL Processes: Develop and maintain Extract, Transform, Load (ETL) processes to handle large volumes of data, ensuring data is processed in a timely and accurate manner.
- Data Governance: Implement data governance and security practices to maintain data quality, compliance, and privacy standards across the data ecosystem.
- Performance Optimization: Identify and address performance bottlenecks in data processing and storage to ensure optimal performance and low-latency data access.
- Version Control: Use Git for version control, ensuring proper documentation and effective collaboration with the team.
- Continuous Integration and Continuous Deployment (CI/CD): Implement CI/CD pipelines to automate the testing and deployment of data pipelines, ensuring the delivery of high-quality data solutions.
- Monitoring and Troubleshooting: Monitor data pipelines and infrastructure to proactively identify issues and perform troubleshooting to ensure data availability and reliability.
- Documentation: Maintain comprehensive documentation of data processes, pipelines, and infrastructure to facilitate knowledge sharing and maintain system understanding.
Benefits
Enroute is committed to providing equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws. This policy applies to all terms and conditions of employment, including recruiting, hiring, placement, promotion, termination, layoff, recall, transfer, leaves of absence, compensation, and training.