Vivo
Arlington, Georgia, United States
Job Type: Contract
We are looking for a skilled Data Engineer to design, develop, and maintain scalable data pipelines and infrastructure. The ideal candidate will be responsible for collecting, storing, processing, and analyzing large datasets to support business intelligence and analytics needs. You will work closely with data scientists, analysts, and software engineers to ensure efficient data flow and accessibility.
Key Responsibilities:
- Design, develop, and optimize data pipelines, ETL processes, and data warehouses.
- Build and maintain real-time and batch data processing solutions.
- Develop and maintain data models, schemas, and databases for structured and unstructured data.
- Implement and maintain data integration solutions across multiple systems.
- Monitor and troubleshoot data pipelines to ensure reliability, accuracy, and efficiency.
- Collaborate with data scientists and analysts to support machine learning models and business intelligence.
- Ensure data quality, governance, and security compliance.
- Optimize and improve database performance and storage solutions.
- Work with big data technologies, cloud services (AWS, GCP, Azure), and distributed systems.
- Automate data workflows and CI/CD pipelines for data processing.
Required Skills & Qualifications:
- Bachelor’s/Master’s degree in Computer Science, Data Engineering, or a related field.
- 3+ years of experience in data engineering, data architecture, or related fields.
- Strong experience with SQL and NoSQL databases (PostgreSQL, MySQL, MongoDB, etc.).
- Expertise in ETL/ELT tools (Apache Airflow, Talend, dbt, etc.).
- Hands-on experience with Big Data technologies (Hadoop, Spark, Kafka, Flink, etc.).
- Proficiency in programming languages (Python, Scala, Java).
- Familiarity with cloud data services (AWS Redshift, Google BigQuery, Azure Synapse, etc.).
- Strong understanding of data modeling, warehousing, and data lake architectures.
- Experience with APIs and data streaming frameworks.
- Knowledge of containerization and orchestration tools (Docker, Kubernetes).
- Strong problem-solving skills and ability to work in an agile environment.
Preferred Skills:
- Experience in machine learning pipelines and MLOps.
- Knowledge of Graph databases (Neo4j, Amazon Neptune).
- Familiarity with CI/CD for data infrastructure.
- Hands-on experience with data governance and security frameworks.
- Experience working with real-time analytics and business intelligence tools (Tableau, Power BI).
Benefits & Perks:
- Competitive salary and performance-based bonuses.
- Health, dental, and vision insurance.
- Flexible working hours and remote work options.
- Learning and career development opportunities.
- Generous paid time off and parental leave.
- Employee wellness programs and team-building activities.