{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/ir-cameras"},"x-facet":{"type":"skill","slug":"ir-cameras","display":"Ir Cameras","count":4},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_126e36d8-668"},"title":"Perception Engineering Intern","description":"<p>We are seeking a perception engineer with a strong background in computer vision to join our rapidly growing team in Costa Mesa, CA. In this role, you will be at the forefront of developing advanced perception systems for complex autonomous aerial platforms.</p>\n<p>Your expertise in computer vision algorithms, combined with your understanding of robotics principles, will be crucial in solving a wide variety of challenges involving visual perception, SLAM, motion planning, controls, and state estimation. This role requires not only technical expertise in computer vision and robotics but also the ability to make pragmatic engineering tradeoffs, considering the unique constraints of aerial platforms.</p>\n<p>Your work will directly contribute to the seamless integration of Anduril&#39;s products, achieving critical outcomes in autonomous operations. This position demands strong systems-level knowledge and experience, as you&#39;ll be working on the intersection of computer vision, robotics, and autonomous systems.</p>\n<p>If you are passionate about pushing the boundaries of computer vision in robotics, possess a &#39;Whatever It Takes&#39; mindset, and can execute in an expedient, scalable, and pragmatic way while keeping the mission top-of-mind and making sound engineering decisions, then this role is for you.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Work at the intersection of 3D perception and computer vision, developing robust algorithms that power real-time decision-making for autonomous aerial systems.</li>\n</ul>\n<ul>\n<li>Develop and implement advanced structure from motion and SLAM algorithms to create accurate 3D models from multiple camera inputs in real-time.</li>\n</ul>\n<ul>\n<li>Integrate perception outputs with path planning algorithms to enable autonomous navigation in complex, unstructured environments.</li>\n</ul>\n<ul>\n<li>Design experiments, data collection efforts, and curate training/evaluation sets to develop insights for both internal purposes and customers.</li>\n</ul>\n<ul>\n<li>Collaborate closely with robotics, software, and hardware teams to integrate perception algorithms into autonomous aerial systems.</li>\n</ul>\n<ul>\n<li>Work with vendors and government stakeholders to advance the state-of-the-art in perception and world modeling for autonomous aerial systems.</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li>BS in Robotics, Computer Science, Mechatronics, Electrical Engineering, Mechanical Engineering, or related field.</li>\n</ul>\n<ul>\n<li>Strong knowledge of 3D computer vision concepts, including multi-view geometry, camera models, photogrammetry, depth estimation, and 3D reconstruction techniques.</li>\n</ul>\n<ul>\n<li>Fluency in standard domain libraries (numpy, opencv, pytorch, etc).</li>\n</ul>\n<ul>\n<li>Proven understanding of data structures, algorithms, concurrency, and code optimization.</li>\n</ul>\n<ul>\n<li>Experience working with Python, PyTorch, or C++ programming languages.</li>\n</ul>\n<ul>\n<li>Experience deploying software to end customers, internal or external.</li>\n</ul>\n<ul>\n<li>Must be willing to travel 25%.</li>\n</ul>\n<ul>\n<li>Eligible to obtain an active U.S. Secret security clearance.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>MS or PhD in Robotics, Computer Science, Engineering, or related field.</li>\n</ul>\n<ul>\n<li>Experience with perception systems for aerial robotics or other highly dynamic platforms.</li>\n</ul>\n<ul>\n<li>Experience with real-world sensor integrations, including LiDAR, RGB-D cameras, IR cameras, stereo cameras, or TOF cameras.</li>\n</ul>\n<ul>\n<li>Experience with GPU / CUDA programming for accelerated computer vision processing.</li>\n</ul>\n<ul>\n<li>Knowledge of path planning algorithms and their integration with perception systems in dynamic environments.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_126e36d8-668","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/4830032007","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"internship","x-salary-range":null,"x-skills-required":["computer vision","robotics","Python","PyTorch","C++","numpy","opencv","data structures","algorithms","concurrency","code optimization"],"x-skills-preferred":["perception systems","aerial robotics","LiDAR","RGB-D cameras","IR cameras","stereo cameras","TOF cameras","GPU","CUDA"],"datePosted":"2026-04-18T15:48:07.380Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"computer vision, robotics, Python, PyTorch, C++, numpy, opencv, data structures, algorithms, concurrency, code optimization, perception systems, aerial robotics, LiDAR, RGB-D cameras, IR cameras, stereo cameras, TOF cameras, GPU, CUDA"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1044b51e-cc6"},"title":"Senior Manager, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs. The role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land. Our Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Lead teams across autonomy, integration, and testing by aligning technical efforts, resolving cross-functional challenges, and driving mission-focused execution.</li>\n<li>Develop advanced perception algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks by integrating data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques.</li>\n<li>Develop state estimation capabilities by designing and refining algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs.</li>\n<li>Analyze and utilize sensor ICDs to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance by tuning and evaluating perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration by working closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings by leveraging synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams to ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing by contributing novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement – Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience.</li>\n<li>Typically requires a minimum of 10 years of related experience with a Bachelor’s degree; or 9 years and a Master’s degree; or 7 years with a PhD; or equivalent work experience.</li>\n<li>7+ years of experience in Unmanned Systems programs in the DoD or applied R&amp;D.</li>\n<li>2+ years of people leadership experience.</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models.</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches.</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications.</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs.</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams.</li>\n<li>Ability to obtain a SECRET clearance.</li>\n</ul>\n<p><strong>Preferences:</strong></p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems.</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks.</li>\n<li>Experience deploying perception software on SWaP-constrained platforms.</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments.</li>\n<li>Understanding of sensing challenges in denied or degraded conditions.</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1044b51e-cc6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/cebc0dd3-ffbf-4013-a2ad-ae32732cabd3","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$229,233 - $343,849 a year","x-skills-required":["BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree","10+ years of related experience","7+ years of experience in Unmanned Systems programs in the DoD or applied R&D","2+ years of people leadership experience","Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models","Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches","Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications","Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs","Proficiency with version control, debugging, and test-driven development in cross-functional teams","Ability to obtain a SECRET clearance"],"x-skills-preferred":["Hands-on integration or algorithm development with airborne sensing systems","Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks","Experience deploying perception software on SWaP-constrained platforms","Familiarity with validating perception systems during flight test events or operational environments","Understanding of sensing challenges in denied or degraded conditions","Exposure to perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:04:16.670Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, DC / San Diego, California / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, 10+ years of related experience, 7+ years of experience in Unmanned Systems programs in the DoD or applied R&D, 2+ years of people leadership experience, Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models, Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches, Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications, Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs, Proficiency with version control, debugging, and test-driven development in cross-functional teams, Ability to obtain a SECRET clearance, Hands-on integration or algorithm development with airborne sensing systems, Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks, Experience deploying perception software on SWaP-constrained platforms, Familiarity with validating perception systems during flight test events or operational environments, Understanding of sensing challenges in denied or degraded conditions, Exposure to perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":229233,"maxValue":343849,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_841c78ea-841"},"title":"Senior Engineer, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs.\nThe role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.\nImplement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.\nDevelop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.\nAnalyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.\nOptimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.\nSupport autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.\nValidate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.\nCollaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.\nDrive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_841c78ea-841","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/d6f1d906-5c1e-4640-87f3-3e31e1b45fa6","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 - $240,000 a year","x-skills-required":["BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience","Typically requires a minimum of 5 years of related experience with a Bachelor’s degree; or 4 years and a Master’s degree; or 2 years with a PhD; or equivalent work experience","Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models","Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches","Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications","Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs","Proficiency with version control, debugging, and test-driven development in cross-functional teams","Ability to obtain a SECRET clearance"],"x-skills-preferred":["Hands-on integration or algorithm development with airborne sensing systems","Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks","Experience deploying perception software on SWaP-constrained platforms","Familiarity with validating perception systems during flight test events or operational environments","Understanding of sensing challenges in denied or degraded conditions","Exposure to perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:03:46.950Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego, California / Washington, DC / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience, Typically requires a minimum of 5 years of related experience with a Bachelor’s degree; or 4 years and a Master’s degree; or 2 years with a PhD; or equivalent work experience, Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models, Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches, Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications, Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs, Proficiency with version control, debugging, and test-driven development in cross-functional teams, Ability to obtain a SECRET clearance, Hands-on integration or algorithm development with airborne sensing systems, Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks, Experience deploying perception software on SWaP-constrained platforms, Familiarity with validating perception systems during flight test events or operational environments, Understanding of sensing challenges in denied or degraded conditions, Exposure to perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bed4759c-578"},"title":"Staff Engineer, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs. The role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land. Our Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.</li>\n<li>Develop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.</li>\n<li>Analyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement , Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p><strong>Required Qualifications:</strong></p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience</li>\n<li>Typically requires a minimum of 7 years of related experience with a Bachelor’s degree; or 5 years and a Master’s degree; or 4 years with a PhD; or equivalent work experience</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams</li>\n<li>Ability to obtain a SECRET clearance</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks</li>\n<li>Experience deploying perception software on SWaP-constrained platforms</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments</li>\n<li>Understanding of sensing challenges in denied or degraded conditions</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms</li>\n</ul>\n<p>$182,720 - $274,080 a year</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bed4759c-578","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/8739c509-b6ea-4640-bcc1-c8b5b1de31b2","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$182,720 - $274,080 a year","x-skills-required":["real-time object detection","sensor fusion","state estimation algorithms","EO/IR cameras","radars","IMUs","Kalman Filters","multi-target tracking","deep learning-based detection models","probabilistic or rule-based approaches","SLAM","visual-inertial odometry","sensor-fused localization","Interface Control Documents","hardware integration specs","version control","debugging","test-driven development"],"x-skills-preferred":["hands-on integration or algorithm development with airborne sensing systems","ML frameworks such as PyTorch or Tensorflow","vision-based object detection or classification tasks","SWaP-constrained platforms","validating perception systems during flight test events or operational environments","sensing challenges in denied or degraded conditions","perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:02:45.901Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego, California / Washington, DC / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"real-time object detection, sensor fusion, state estimation algorithms, EO/IR cameras, radars, IMUs, Kalman Filters, multi-target tracking, deep learning-based detection models, probabilistic or rule-based approaches, SLAM, visual-inertial odometry, sensor-fused localization, Interface Control Documents, hardware integration specs, version control, debugging, test-driven development, hands-on integration or algorithm development with airborne sensing systems, ML frameworks such as PyTorch or Tensorflow, vision-based object detection or classification tasks, SWaP-constrained platforms, validating perception systems during flight test events or operational environments, sensing challenges in denied or degraded conditions, perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":182720,"maxValue":274080,"unitText":"YEAR"}}}]}