Data Engineer, GCP
Richmond, VA
Job Type: Contract to Hire | Recruiter: Holly Bajar | Phone: 617.337.4251
Location: Virginia, USA (Remote – candidates must be in EST or CST)
Length: 6+ months
*Candidates must be a United States Citizen or United States Permanent Resident
Job Description:
Our client is seeking an experienced and highly skilled Data Engineer to join their team and support the design, implementation, and maintenance of robust data pipelines and infrastructure within a cloud environment. The ideal candidate will possess expertise in Google Cloud Platform (GCP) technologies, including BigQuery, BigLake, and Gemini, to enable advanced analytics and business intelligence. You will play a crucial role in managing large datasets, optimizing data workflows, and utilizing cloud tools to deliver impactful insights that support business decision-making.
Essential Duties and Responsibilities (but not limited to):
- Design and implement scalable, efficient data architectures and manage ETL/ELT pipelines using GCP tools such as BigQuery, BigLake, and Dataflow.
- Integrate structured and unstructured data sources to build unified platforms and ensure seamless data flow across systems.
- Leverage BigQuery for large-scale data analysis and business intelligence.
- Write and optimize SQL queries and automate data pipelines for regular and real-time processing.
- Apply strategies such as partitioning and clustering to optimize data storage and query performance.
- Develop and implement data integration strategies using GCP services (i.e., Cloud Pub/Sub and Dataflow).
- Work with machine learning models, particularly through Gemini and Vertex AI, to integrate predictive analytics and real-time insights into data pipelines.
- Implement data governance and security practices, including setting up access controls, encryption, and auditing procedures to ensure compliance and protect sensitive information.
- Collaborate with cross-functional teams to deliver data solutions that meet business needs. Provide mentorship and technical guidance to junior team members.
- Bachelor’s degree in Computer Science, Engineering, Data Science, or a related field (or equivalent experience).
- At least 3 years of experience working with Google Cloud Platform (GCP), particularly with BigQuery, BigLake, and other GCP tools.
- Strong proficiency in SQL for querying, transforming, and optimizing large datasets.
- Experience designing and managing ETL/ELT pipelines using GCP services such as Dataflow, Dataproc, or Cloud Composer.
- Familiarity with BigLake for integrating structured and unstructured data sources.
- Understanding of cloud data storage and processing concepts such as data lakes, data warehouses, and object storage.
- Experience working with machine learning models, especially in integration with GCP services like Gemini or Vertex AI.
- Experience with advanced machine learning algorithms and model deployment, particularly on cloud platforms is preferred.
- Familiarity with orchestration tools (i.e., Apache Beam, Apache Airflow) and CI/CD processes for data pipeline automation is a plus.
- Strong understanding of data governance, security, privacy, and cloud resource cost optimization is a plus.
- Candidates must be a United States Citizen or United States Permanent Resident.
Come see why DPC has achieved:
- 4.9/5 Star Glassdoor rating and the only staffing company (< 1000 employees) to be voted in the national Top 10 ‘Employee’s Choice - Best Places to Work’ by Glassdoor.
- Voted ‘Best Staffing Firm to Temp/Contract For’ seven times by Staffing Industry Analysts as well as a ‘Best Company to Work For’ by Forbes, Fortune and Inc. magazine.
www.LoveYourJob.com