Apptronik is building robots for the real world to improve human quality of life and to help solve the ever-increasing labor shortage problem. Our team has been building some of the most advanced robots on the planet for years, dating back to the DARPA Robotics Challenge. We apply our expertise across the full robotics stack to some of the most important and impactful problems our society faces, and expect our products and technology to change the world for the better. We value passion, creativity, and collaboration to help us overcome existing technological barriers in the industry to create truly innovative products.
You will join a team developing state-of-the-art general-purpose robots designed to operate in human spaces and with human tools. It is designed to work alongside humans, mobilize to human spaces, and manipulate the world around it.
As a Senior Perception Learning Engineer, you will lead research and development of advanced perception systems that empower Apptronik’s humanoid robots to understand and interact with complex human environments. Your work will focus on cutting-edge research in perception, SLAM, object detection, world modeling, and multi-sensor fusion, creating the foundation for robust autonomy in real-world settings.
You will design and optimize deep learning models for real-time detection, tracking, segmentation, and scene understanding while architecting scalable pipelines for training, evaluation, and deployment. You will also integrate data from multiple modalities—Cameras, LiDAR, depth sensors, and IMUs—into unified world models that support navigation, manipulation, safety and human-robot interaction.
This role requires balancing research innovation with practical engineering to deliver deployable, high-performance perception stacks. You will collaborate across Reinforcement learning teams, Platform software team and systems teams, mentor junior engineers, and contribute to shaping Apptronik’s long-term perception and autonomy roadmap. Your work will directly accelerate the development of humanoid robots that can safely operate in human spaces, adapt to dynamic environments, and extend human capability.
Responsibilities- Lead the design, development, and optimization of perception pipelines for humanoid robots, including object detection, tracking, segmentation, pose estimation, and scene understanding.
- Develop multi-sensor fusion frameworks that integrate cameras, LiDAR, depth sensors, and IMUs for robust real-time perception in dynamic human-centered environments.
- Architect and maintain scalable data pipelines, training infrastructure, and inference frameworks to accelerate model development, evaluation, and deployment.
- Drive research and deployment of deep learning models optimized for humanoid locomotion, manipulation, and human-robot interaction.
- Implement performance profiling, regression testing, and telemetry systems to ensure perception modules meet strict latency, accuracy, and reliability requirements on edge devices.
- Collaborate with planning, control, and hardware teams to define perception-to-action interfaces, ensuring real-time compatibility with locomotion and manipulation pipelines.
- Guide the integration of synthetic data (e.g., simulation frameworks like IsaacSim) with real-world datasets to enhance model generalization and robustness.
- Mentor junior engineers and contribute to best practices in code quality, model versioning, reproducibility, and deployment.
- MS/PhD in Computer Science, Robotics, Computer Engineering, or related field.
- 3-5+ years of experience building and deploying perception systems for robotics, autonomous vehicles, or real-time vision applications.
- Strong background in deep learning for computer vision, with practical expertise in detection, segmentation, multi-object tracking, and 3D perception.
- Hands-on experience with modern AI frameworks (PyTorch, JAX, TensorFlow) and computer vision / multi-modal libraries such as OpenCV, Detectron2, YOLO, and foundation models for perception and language (e.g., SAM, CLIP, DINOv2, Flamingo)
- Proficiency in Python and modern C++, with strong software engineering fundamentals (version control, testing, CI/CD).
- Deep understanding of 3D geometry, camera models, and probabilistic estimation (EKF/UKF, SLAM, VIO).
- Experience deploying optimized models on edge hardware (GPU/NPU/embedded platforms) under compute, latency, and thermal constraints.
- Track record of shipping ML/Perception systems from R&D into production robotics platforms.
- Experience with humanoid robots, bipedal locomotion, and manipulation tasks.
- Strong classical computer vision skills (geometry-based methods, feature extraction) complementing deep learning approaches.
- Expertise in model acceleration, quantization, or compression (TensorRT, ONNX Runtime).
- Familiarity with real-time frameworks and middleware such as ROS 2, GStreamer, or zero-copy pipelines.
- Knowledge of synthetic data generation and domain adaptation techniques for training perception models.
- Contributions to open-source robotics or vision software stacks.
PHYSICAL REQUIREMENTS
- [Prolonged periods of sitting at a desk and working on a computer
- Must be able to lift 15 pounds at times
- Vision to read printed materials and a computer screen
- Hearing and speech to communicate]
*This is a direct hire. Please, no outside Agency solicitations.
Apptronik provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.
Top Skills
Similar Jobs at Apptronik
What you need to know about the Charlotte Tech Scene
Key Facts About Charlotte Tech
- Number of Tech Workers: 90,859; 6.5% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Lowe’s, Bank of America, TIAA, Microsoft, Honeywell
- Key Industries: Fintech, artificial intelligence, cybersecurity, cloud computing, e-commerce
- Funding Landscape: $3.1 billion in venture capital funding in 2024 (CED)
- Notable Investors: Microsoft, Google, Falfurrias Management Partners, RevTech Labs Foundation
- Research Centers and Universities: University of North Carolina at Charlotte, Northeastern University, North Carolina Research Campus