A
AI

AI ML Engineer

ADN Group · San Francisco, CA

Actively hiring Posted 7 days ago

Role overview

AI/ML Engineer (AI Voice & Social Product) - w/ Equity

Location: San Francisco, CA (Onsite 5 days a week)

Type: Full-Time

Job Overview

Own ML systems for a voice-AI product driving accurate match-making and continuous improvement. Collaborate on data pipelines, model training, evaluation, and deployment with emphasis on efficiency and low-latency. Key responsibilities: Design multi-stage retrieval and re-ranking for personalization; manage data pipelines ensuring reproducibility; train and fine-tune LLMs; run offline and online evaluations; set latency and cost targets for inference services.

Required skills: 10+ years in …

What we're looking for

Known is building a voice-AI product that powers curated introductions, agentic scheduling, and post-date feedback. You will own the ML systems that make matches feel accurate and improve every week—from data and features to training, evaluation, and low-latency inference—working closely with platform and product.

Our stack: Python, PyTorch, Hugging Face, OpenAI/Anthropic APIs, embeddings and vector search (pgvector/Pinecone/FAISS), Postgres + warehouse for analytics, Airflow/Prefect/dbt for pipelines, online experimentation/A/B testing, observability for models and services on AWS (S3, ECS/Kubernetes, Lambda). CI/CD with GitHub Actions.

• Design and ship multi-stage retrieval + re-ranking for compatibility scoring, search, and personalization.

• Build and maintain data/feature pipelines for training, evaluation, and reporting; ensure reproducibility and data quality.

• Train, fine-tune, or prompt LLM/encoder models; manage model versioning, rollout, and rollback.

• Run offline evaluation (e.g., AUC, NDCG, MAP) and online experiments to measure real user impact.

• Stand up inference services with tight p95 latency and cost targets; add caching, batching, and fallback strategies.

• Implement safety/guardrails and monitoring for drift, bias, and failure modes; define model SLOs and alerts.

• Collaborate with infra/platform to productionize models and with product/design to turn signals from voice/text into better matches.

• Document decisions, write lightweight runbooks, and share dashboards that track match quality and model health.

We are hiring a founding-caliber Infrastructure / Platform Engineer who has owned production cloud environments and data platforms in high-growth settings. You will set the golden paths for services, data, and model delivery, and you are comfortable working on-site in San Francisco five days a week.

• 4 to 10+ years in infrastructure, platform, or data engineering with real ownership of uptime, performance, and security.

• Expert with AWS and Infrastructure-as-Code (Terraform, Pulumi, or CloudFormation).

• Strong proficiency in Python or TypeScript, plus tooling/scripting (Bash/YAML).

Must-Have Requirements

✓ Must be authorized to work in the U.S. without future visa sponsorship.

✓ Able to work onsite in San Francisco, CA five days per week.

✓ 3+ years in applied ML focused on ranking, recommendations, or search in production.

✓ Strong Python; experience with PyTorch or TensorFlow (Hugging Face a plus).

✓ Hands-on with embeddings and vector search (pgvector, FAISS, Pinecone, or Weaviate).

✓ Proven experience taking models from notebook to production: packaging, APIs, CI/CD, canary/rollback, monitoring.

✓ Data pipelines for training and evaluation (e.g., Airflow, Prefect, Dagster, or dbt) and sound data-quality checks.

Benefits

• Containers and orchestration experience (Docker, Kubernetes or ECS) and CI/CD pipelines you designed and ran.

• Proven ability to design and operate data pipelines and distributed systems for both batch and low-latency use cases.

• PostgreSQL at scale, ideally with pgvector/embeddings exposure for ML-adjacent workloads.

• Strong observability practices: metrics, tracing, alerting, incident management, and SLOs.

• Excellent collaboration with AI/ML and product teams; clear communication of tradeoffs and risk.

• Work authorization in the U.S. and willingness to be on-site five days a week in San Francisco.

• Experience supporting model training and inference pipelines, feature stores, or evaluation loops.

• Prior work with streaming voice, low-latency systems, or recommendation/retrieval stacks.

Tags & focus areas

Used for matching and alerts on DevFound
Fulltime Ai Engineer Machine Learning