Job Title - Data MLOps
Description - Data MLOps engineer resource will be member of VWGoA Enterprise Data and Advanced Analytics team, working with cross-functional teams to leverage data and advanced analytical techniques and deliver data-driven solutions within an organization. This work entails supporting data ingestion processes and problem solving in a variety of domains, including: business process optimization; and conducting research and analysis on new technologies and business methodologies.
Functions/Responsibility Performed:
• Design & Develop Data Pipelines:
o Orchestrate complex data pipelines using data integration tools like Informatica and Python, ensuring seamless data flow from various sources.
o Leverage GCP Dataflow, Cloud Functions, and other cloud technologies to build scalable and resilient data ingestion and processing pipelines.
o Implement robust CI/CD workflows using GitHub Actions and Argo CD to automate pipeline deployments and ensure consistency.
o Monitor and manage production solutions. Optimize and fine-tune models for performance, accuracy, and scalability.
o Document best practices and quality standards to be adherence during development of data science solutions.
o Conduct review and provide feedback on data science work applications.
• Manage & Analyze Data:
o Work with diverse data sources, including relational databases (Oracle, SQL Server, MySQL, Postgres, Snowflake), big data platforms (Hadoop, Parquet files, BigQuery, Big Lake managed Iceberg), and streaming data (Kafka, GCP Dataflow/Proc).
o Employ powerful compute engines like Hive, Impala, and Spark to analyze massive datasets and derive valuable insights.
• Deliver Actionable Insights:
o Collaborate with business stakeholders to understand their challenges and requirements.
o Translate business problems into analytical frameworks and identify opportunities to address complex problems.
o Build APIs and user-friendly interfaces to present data results and empower informed decision-making.
• Drive Machine Learning Innovation:
o Explore and implement Vertex AI models to generate quick insights and support business requirements.
• Stay at the Forefront:
o Continuously learn and adapt to emerging data technologies and best practices.
o Contribute to the ongoing improvement of data infrastructure and processes.
Functions/Responsibility Performed:
• Design & Develop Data Pipelines:
o Orchestrate complex data pipelines using data integration tools like Informatica and Python, ensuring seamless data flow from various sources.
o Leverage GCP Dataflow, Cloud Functions, and other cloud technologies to build scalable and resilient data ingestion and processing pipelines.
o Implement robust CI/CD workflows using GitHub Actions and Argo CD to automate pipeline deployments and ensure consistency.
o Monitor and manage production solutions. Optimize and fine-tune models for performance, accuracy, and scalability.
o Document best practices and quality standards to be adherence during development of data science solutions.
o Conduct review and provide feedback on data science work applications.
• Manage & Analyze Data:
o Work with diverse data sources, including relational databases (Oracle, SQL Server, MySQL, Postgres, Snowflake), big data platforms (Hadoop, Parquet files, BigQuery, Big Lake managed Iceberg), and streaming data (Kafka, GCP Dataflow/Proc).
o Employ powerful compute engines like Hive, Impala, and Spark to analyze massive datasets and derive valuable insights.
• Deliver Actionable Insights:
o Collaborate with business stakeholders to understand their challenges and requirements.
o Translate business problems into analytical frameworks and identify opportunities to address complex problems.
o Build APIs and user-friendly interfaces to present data results and empower informed decision-making.
• Drive Machine Learning Innovation:
o Explore and implement Vertex AI models to generate quick insights and support business requirements.
• Stay at the Forefront:
o Continuously learn and adapt to emerging data technologies and best practices.
o Contribute to the ongoing improvement of data infrastructure and processes.