Tallo logoTallo logo

Software Engineer - Machine Learning III

Job

DGN Technologies

Mountain View, CA (In Person)

Full-Time

Posted 3 days ago (Updated 12 hours ago) • Actively hiring

Expires 6/13/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
100
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Job Title:
Software Engineer -
Machine Learning III Duties:
Machine Learning Engineer, Prompt Safety & Agent Security Lab Summary The Developer Quality Innovation Lab at Samsung Research America builds the automation and tooling that powers data acquisition, safety, and evaluation for Samsung's mobile platform products. Our systems collect, curate, augment data and develop intelligent solution to protect models that fuels the foundation models and AI features shipping across Galaxy devices - and operate the evaluation pipelines that gate their quality before and after launch. We work closely with modeling, device, and product teams to close the loop from on-device signals and user feedback back into training data, faster and at higher quality. Position Summary We are looking for an experienced Machine Learning Engineer to lead the development of prompt injection and prompt safety models that protect Samsung's downstream agentic AI systems across phone, cloud, and XR/AR. You will design, train, and deploy classifier and guardrail models (both cloud-based and hybrid on-device) that screen agent inputs and outputs for injection attacks, unsafe content, and policy violations. A core part of the role is post-training these models with RLHF, DPO, and related optimization techniques to push detection accuracy and false-positive rates beyond what off-the-shelf solutions provide. Role and Responsibilities 1. Design and train prompt injection detection models and prompt safety classifiers that operate on both inputs to and outputs from Samsung's agentic AI systems. 1. Build hybrid deployment pipelines that split safety inference between on-device (phone, XR/AR) and cloud, optimizing for latency, privacy, and detection coverage. 1. Apply post-training techniques (e.g. RLHF, reward modeling, policy optimization) to optimize guardrail model performance, calibration, and robustness against adaptive adversaries. 1. Curate and generate adversarial training data: direct and indirect prompt injections, jailbreaks, tool-use exploits, and unsafe-output cases drawn from red-teaming and production signals. 1. Build evaluation harnesses that measure attack success rate, false-positive rate, latency, and on-device footprint across model iterations and threat categories. 1. Partner with agent, device, and platform teams to integrate safety models into mobile-use agents, XR/AR assistants, and cloud agentic workflows, and to close the loop from production incidents back into training data. 1. Work cross-functionally with security researchers, modeling teams, and product engineers; document methods and, where appropriate, contribute to patents and publications. Required Qualifications 1. M.S. or Ph.D. in Computer Science, Machine Learning, Electrical Engineering, or a related field; or B.S. with equivalent industry experience. 1. 3+ years of industry experience in ML engineering or applied AI research, with demonstrated ownership of production ML systems. 1. 2+ years of industry experience in software engineering. 1. Strong proficiency in Python and PyTorch (or JAX/TensorFlow), with solid software engineering fundamentals (version control, testing, and reproducible experimentation). 1. Hands-on experience post-training LLMs with RLHF, DPO, RLAIF, or reward modeling including reward design, preference data curation, and training stability. 1. Hands-on experience training and deploying classifier or guardrail models for safety, content moderation, abuse detection, or adversarial robustness. 1. Familiarity with prompt injection, jailbreak, and agentic AI threat models, and with distributed training frameworks (DeepSpeed, FSDP, Accelerate). Preferred Qualifications 1. Experience building safety or moderation systems for agentic
AI:
tool-use guardrails, indirect prompt injection defenses, or output filtering for autonomous agents. 1. Experience with red-teaming, adversarial data generation, or automated attack pipelines (e.g.,
GCG, Skills:
PAIR, generator-critic frameworks). 1. Experience with on-device or edge ML deployment (ExecuTorch, Core ML, TFLite, MLC-LLM, vendor NPU toolchains) and model compression (quantization, distillation, pruning) for safety models. 1. Experience with telemetry, logging, or user-facing data systems on mobile, XR/AR, or consumer platforms, including privacy-preserving handling of user data (e.g., anonymization, on-device processing, federated approaches). 1. Publications at top-tier ML/NLP/security venues (Neur
IPS, ICML, ICLR, ACL, EMNLP, USENIX
Security, IEEE S&P), patents, or open-source contributions in the safety, alignment, or AI security space.
Keywords:
Education:
Skills and Experience:
Required Skills:
MACHINE LEARNING ENGINEERING APPLIED AI RESEARCH SOFTWARE ENGINEERING PYTHON PYTORCH
Additional Skills:
JAX TENSORFLOW VERSION CONTROL TESTING REPRODUCIBLE EXPERIMENTATION POST-TRAINING LLMS RLHF DPO RLAIF REWARD MODELING REWARD DESIGN PREFERENCE DATA CURATION TRAINING STABILITY CLASSIFIER TRAINING GUARDRAIL MODEL TRAINING SAFETY MODEL DEPLOYMENT CONTENT MODERATION ABUSE DETECTION ADVERSARIAL ROBUSTNESS PROMPT INJECTION DETECTION JAILBREAK DETECTION AGENTIC AI THREAT MODELING DISTRIBUTED TRAINING FRAMEWORKS DEEPSPEED FSDP ACCELERATE SAFETY SYSTEM DEVELOPMENT FOR AGENTIC AI TOOL-USE GUARDRAILS INDIRECT PROMPT INJECTION DEFENSES OUTPUT FILTERING FOR AUTONOMOUS AGENTS RED-TEAMING ADVERSARIAL DATA GENERATION AUTOMATED ATTACK PIPELINES GCG PAIR
Languages:
English Read Write Speak Minimum Degree Required:
Master's Degree Patents:
No Publications:
No Veteran Status:
No

Similar remote jobs

Similar jobs in Mountain View, CA

Similar jobs in California