Senior AI Research Engineer, Visual Perception
Description
ANYbotics is a fast-growing tech company dedicated to shaping the future of mobile robotics across multiple industries. You’ll join our Solutions team, where we push the limits of autonomous robotic inspection intelligence. You’ll build cutting-edge AI, machine-learning, and sensing technologies, from early exploration to tackle the toughest real-world deployment challenges. Collaborating with engineers and partners, you’ll assess new technologies, design proof-of-concept experiments, build research prototypes, and swiftly transfer promising ideas to our product. Your work will shape our technology roadmap, guiding long-term strategy and ensuring every solution is robust, reliable, and ready for impact.
Market & Technology ANYbotics transforms industrial plants in the (renewable) energy, process, and utility sector by introducing robotics to a wide range of novel applications that so far were beyond reach. Our mobile robot ANYmal uses legs for extreme mobility in complex environments, camera- and LIDAR-based sensing for full autonomy and obstacle avoidance, to perform jobs and deliver high-quality, consistent inspection results. We develop numerous customized hardware systems, including the entire robotic platform, actuators, sensors, inspection payloads, charging systems, and all related ANYbotics electrical hardware
Responsibilities
- Exploring, developing, and implementing cutting-edge technological advancements in multi modal (vision, audio, thermal, gas,...) machine learning.
- Identifying emerging technologies, conducting in-depth evaluations, and translating these findings into innovative products.
- Developing prototypes and conducting pilot projects to test new technologies in controlled environments.
- Driving collaborations with tech startups, research institutions, and industry leaders to leverage external expertise and accelerate innovation.
- Connecting with the engineering teams for the adoption and integration of new technologies into existing systems and processes, ensuring a seamless transition.
Qualifications
- A PhD in Machine Learning (or a related technical field) OR 4+ years of relevant industrial hands-on experience.
- Excellent at designing and deploying state-of-the-art computer vision and AI perception models including multi-modal approaches.
- Excellent expertise in modern deep learning techniques, including transformer-based architectures, large Vision-language models (VLMs), Video generative models and diffusion algorithms or Vision Language Action models (VLAs) with hands-on experience training and fine-tuning such systems for scene understanding applications.
- Strong track record of implementing and evaluating advanced AI algorithms for perception and scene understanding in real-world scenarios.
- Strong practical experience across the MLOps pipeline—from data collection and curation to training, fine-tuning, and evaluation.
- Strong in fast prototyping, concept validation, and rapid iteration—you build to learn and use experimentation as a core development tool.
- Strong programming skills in Python, with the ability to write clean, and efficient code.
- Experienced in leading R&D initiatives or small cross-functional teams, with a focus on transferring promising ideas into impactful products.
- Strong in self-coordination and project ownership, able to structure your work independently, align with strategic goals, and deliver meaningful outcomes.
- Excellent communication skills, with the ability to effectively share knowledge and insights.
- Experience with large scale machine learning compute systems.
- Demonstrated coding, debugging, and engineering skills in writing modern C++.
- Good familiarity with GPU programming, performance benchmarking and optimization, to deploy on edge or cloud systems