- Страна
- США
Откликайтесь
на вакансии с ИИ

MLOps Engineer — AI/ML Systems & Deployment (TS/SCI Preferred)
Отличная вакансия для опытных инженеров: работа с передовыми технологиями (LLM, CV) в проектах национального значения. Высокий уровень ответственности компенсируется отличным соцпакетом (100% покрытие обучения, 401k) и уникальным опытом работы в оборонном секторе.
Сложность вакансии
Высокая сложность обусловлена необходимостью работы в закрытых средах (TS/SCI) и глубокими знаниями на стыке ML, DevOps и распределенных систем. Требуется опыт не просто в разработке моделей, а в их промышленной эксплуатации в критически важных инфраструктурах.
Анализ зарплаты
Предлагаемая роль MLOps инженера в сфере национальной безопасности США обычно оплачивается выше среднего по рынку из-за требований к уровню допуска (TS/SCI) и дефицита кадров на стыке ML и инфраструктуры. Рыночные оценки для Senior/Lead позиций в Огайо и удаленно для госсектора находятся в диапазоне $150k-$190k.
Сопроводительное письмо
I am writing to express my strong interest in the MLOps Engineer position at Rackner. With a robust background in deploying production-grade ML pipelines and managing Kubernetes-based infrastructure, I am eager to contribute to your mission-critical AI/ML systems supporting Air Force and NASIC-aligned programs. My experience aligns perfectly with your need for an engineer who can bridge the gap between experimental Jupyter notebooks and reliable, containerized production environments.
In my previous roles, I have successfully orchestrated workflows using Kubeflow and Airflow, ensuring model versioning and lineage were maintained to the highest standards. I am particularly drawn to Rackner's focus on reliability and auditability in high-stakes environments. Having worked extensively with Python, PyTorch, and containerization tools like Docker, I am confident in my ability to operationalize complex LLM and computer vision models within your cloud-native infrastructure.
I am excited about the opportunity to bring my technical expertise in MLOps and my commitment to building repeatable, scalable systems to your team. Thank you for considering my application. I look forward to the possibility of discussing how my skills can support Rackner’s national security impact.
Составьте идеальное письмо к вакансии с ИИ-агентом

Откликнитесь в rackner уже сейчас
Присоединяйтесь к Rackner, чтобы перевести ИИ из стадии экспериментов в реальные боевые системы национального масштаба!
Описание вакансии
MLOps Engineer — AI/ML Systems & Deployment (TS/SCI Preferred)
Dayton, OH (On-site Preferred) | Remote Eligible (CAC-Ready Candidates)
Mission Environment | AI/ML Infrastructure | National Security Impact
About the Role
At Rackner, we are building the operational backbone that turns AI/ML capability into real-world mission outcomes. We are seeking an MLOps Engineer to own the lifecycle of AI/ML systems—from experimentation to deployment—within a mission-critical, classified environment supporting Air Force and NASIC-aligned programs.
This is not a research role; This is where models become reliable, deployable, auditable systems.
You will operate at the intersection of:
- Machine learning
- Distributed systems
- Cloud-native infrastructure
…and ensure that AI/ML systems work in the environments where failure is not an option.
What You’ll Do
Own the ML Lifecycle (End-to-End)
- Build and operate production-grade ML pipelines
- Orchestrate workflows using Kubeflow, Airflow, or Argo
- Implement model versioning, lineage, and reproducibility standards
Operationalize AI/ML Systems
- Deploy models into mission environments (including constrained or classified systems)
- Transition workflows from Jupyter experimentation → containerized pipelines → production systems
- Enable both batch and real-time inference architectures
Engineer for Reliability, Not Just Performance
- Design systems for reproducibility, auditability, and stability
- Implement monitoring for:
+ model performance & drift
+ system health & latency
- Use tools like Prometheus, Grafana, and OpenTelemetry
Build Cloud-Native ML Infrastructure
- Deploy and manage Kubernetes-based ML workloads
- Containerize pipelines using Docker / OCI standards
- Scale compute for training and inference workloads
Establish Data Discipline
- Enable data versioning and governance (lakeFS or similar)
- Support feature engineering and dataset preparation pipelines
- Apply metadata standards (e.g., STAC) where applicable
Create Repeatable Systems
- Develop runbooks, playbooks, and deployment standards
- Build systems that can be operated by others; not just understood by you
What You Bring
Core Experience
- Experience deploying ML systems into production environments
- Strong background in Python and ML frameworks (PyTorch, TensorFlow, etc.)
- Hands-on experience with:
+ ML pipeline orchestration tools (Kubeflow, Airflow, Argo)
+ Experiment tracking (MLflow, ClearML)
Infrastructure & Systems
- Experience with Kubernetes and containerized workloads
- Familiarity with CI/CD for ML systems
- Understanding of distributed systems and scalable architectures
ML Application Exposure
- Experience working with:
+ LLMs or transformer-based models
+ computer vision systems (YOLO, Faster R-CNN)
- Focus on deployment and integration, not pure research
Mindset
- Systems thinker who values reliability over novelty
- Comfortable operating in ambiguous, high-stakes environments
- Able to translate experimental work into operational capability
Why This Role Matters (What You Get)
This role is a career accelerator for engineers who want to:
- Move beyond experimentation
+ Own systems that actually get deployed and used
- Operate at the systems level
+ Work across ML, infrastructure, and mission integration
- Build in high-trust environments
+ Where correctness, auditability, and reliability matter
- Develop rare, high-demand expertise
+ MLOps in constrained / classified environments is a differentiated skillset
Shape how AI is operationalized—not just built
Who We Are
Rackner is a software consultancy that builds cloud-native solutions for startups, enterprises, and the public sector. We are an energetic, growing consultancy with a passion for solving big problems across industries.
We enable digital transformation through:
- Distributed systems
- DevSecOps
- AI/ML
- Cloud-native architecture
Our approach is cloud-first, cost-effective, and outcome-driven—focused on delivering real capability, not just code.
Benefits & Perks
- 100% covered certifications & training aligned to your role
- 401(k) with 100% match up to 6%
- Highly competitive PTO
- Comprehensive Medical, Dental, Vision coverage
- Life Insurance + Short & Long-Term Disability
- Home office & equipment plan
- Industry-leading weekly pay schedule
Apply
If you’re an engineer who wants to move from building models → owning systems, we want to talk.
#MLOps #MachineLearning #Kubernetes #AIEngineering #CloudNative #DevSecOps #ArtificialIntelligence #DataEngineering #DefenseTech #NationalSecurity #AIInfrastructure #Hiring #TechCareers
Создайте идеальное резюме с помощью ИИ-агента

Навыки
- Python
- PyTorch
- LLM
- Kubernetes
- Prometheus
- Grafana
- OpenTelemetry
- Computer Vision
- Docker
- Airflow
- TensorFlow
- Kubeflow
- MLflow
- DevSecOps
- Argo
- YOLO
- ClearML
- lakeFS
- STAC
Возможные вопросы на собеседовании
Роль требует перевода моделей из Jupyter в продакшн. Важно понимать, как кандидат обеспечивает воспроизводимость.
Опишите ваш процесс контейнеризации ML-пайплайна: как вы обеспечиваете идентичность среды обучения и инференса?
Вакансия упоминает Kubeflow и Airflow. Нужно оценить опыт работы с оркестраторами.
В каких ситуациях вы предпочтете Kubeflow вместо Airflow для управления жизненным циклом модели и почему?
Работа ведется в закрытых/ограниченных средах, где мониторинг критичен.
Как бы вы организовали мониторинг дрейфа данных (data drift) и производительности модели в изолированном контуре (air-gapped environment)?
Упоминается работа с LLM и CV. Нужно проверить понимание специфики деплоя тяжелых моделей.
С какими основными трудностями вы сталкивались при масштабировании инференса для LLM или тяжелых моделей Computer Vision в Kubernetes?
Вакансия подчеркивает важность 'Data Discipline'.
Расскажите о вашем опыте внедрения версионирования данных (например, через lakeFS или DVC). Как это помогло в аудите системы?
Похожие вакансии
MLOps Engineer (ML pipelines / Kubernetes / Airflow)
Senior Data инженер
Senior MLOps Engineer (Platform Development / LLMOps)
Data Engineer / SAP HANA Developer (Senior)
Data Scientist в RecSys
Data Engineering Team Lead (команда Clickhouse)
1000+ офферов получено
Устали искать работу? Мы найдём её за вас
Quick Offer улучшит ваше резюме, подберёт лучшие вакансии и откликнется за вас. Результат — в 3 раза больше приглашений на собеседования и никакой рутины!
- Страна
- США