PROCESSING APPLICATION
Hold tight! We’re comparing your resume to the job requirements…

ARE YOU SURE YOU WANT TO APPLY TO THIS JOB?
Based on your Resume, it doesn't look like you meet the requirements from the employer. You can still apply if you think you’re a fit.
Job Requirements of ML Data Engineer:
-
Employment Type:
Full-Time
-
Location:
San Francisco, CA (Onsite)
Do you meet the requirements for this job?
ML Data Engineer
Job Summary:
We are seeking a highly skilled Data/ML Engineer to design, build, and optimize data pipelines and machine learning infrastructure. You will work closely with data scientists, analysts, and software engineers to enable scalable machine learning models and real-time data processing. This role requires expertise in big data processing, cloud technologies, and MLOps best practices to support AI-driven applications.
Key Responsibilities:
-
Design, develop, and maintain ETL/ELT pipelines for structured and unstructured data.
-
Implement scalable data architectures using cloud platforms (AWS, GCP, Azure).
-
Develop and optimize machine learning pipelines for training, validation, and deployment.
-
Work closely with data scientists to productionize ML models using MLflow, TensorFlow, PyTorch, or Scikit-learn.
-
Implement MLOps best practices, including CI/CD pipelines for model deployment and monitoring.
-
Optimize data storage and retrieval using data lakes, warehouses (Snowflake, Redshift, BigQuery), and NoSQL databases.
-
Develop real-time data streaming solutions using Kafka, Kinesis, or Apache Flink.
-
Ensure data quality, governance, and compliance with industry standards.
-
Automate workflows using Airflow, Prefect, or Dagster.
-
Monitor model performance and ensure retraining pipelines are in place.
Required Skills & Qualifications:
-
3-7 years of experience in Data Engineering, ML Engineering, or a related field.
-
Strong programming skills in Python and SQL (experience with Scala or Java is a plus).
-
Hands-on experience with big data processing frameworks (Spark, Hadoop, Dask).
-
Expertise in machine learning frameworks (TensorFlow, PyTorch, Scikit-learn).
-
Experience with containerization (Docker, Kubernetes) for model deployment.
-
Familiarity with feature engineering, feature stores, and vector databases (Feast, Pinecone).
-
Strong understanding of data pipelines, batch & streaming data processing.
-
Knowledge of MLOps tools (Kubeflow, MLflow, Sagemaker, Vertex AI).
-
Proficiency in cloud services such as AWS (S3, Lambda, SageMaker), GCP (BigQuery, Vertex AI), or Azure (Synapse, ML Studio).
-
Experience with monitoring and logging tools (Prometheus, Grafana, ELK Stack).
Nice to Have:
-
Experience working in Retail, Finance, Healthcare, or E-commerce domains.
-
Exposure to A/B testing, recommendation systems, or NLP applications.
-
Understanding of data privacy regulations (GDPR, CCPA).