{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/operational-environments"},"x-facet":{"type":"skill","slug":"operational-environments","display":"Operational Environments","count":10},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8c83efd9-b0f"},"title":"Cell and Gene Therapy (CGT) Quality Operations Subject Matter Expert","description":"<p>At Bayer, we&#39;re seeking a Cell and Gene Therapy (CGT) Quality Operations Subject Matter Expert to play a pivotal role in advancing the CGT Quality Assurance (QA) program at our Berkeley site.</p>\n<p>This position involves active participation in the daily project and operational tasks within the CGT manufacturing facilities. Primarily focused on on-the-floor Quality Operations, the SME is responsible for enhancing the team&#39;s understanding of Cell and Gene Therapy quality requirements.</p>\n<p>The individual in this role must possess the ability to work independently, demonstrating excellent judgment and the capability to effectively engage with a diverse range of key stakeholders.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Fully own the Quality Operational tasks required to ensure compliant and safe execution of CGT operational activities in the CGT MOD facility.</li>\n<li>Oversee quality assurance processes for CGT products from development to commercial launch.</li>\n<li>Engage with cross-functional teams to ensure compliance with Global Processes and Regulatory standards.</li>\n<li>Manage risks associated with CGT product development and manufacturing, including finding risk-based and phase-appropriate solutions when issues arise.</li>\n<li>Foster a culture of open communication and collaboration among team members and stakeholders, with an aptitude to being able to have tough conversations when Quality decisions may not be favored across stakeholder groups.</li>\n<li>Provide coaching and mentorship to team members, promoting their professional growth and ensuring alignment to the Quality decisions made.</li>\n<li>Develop and implement strategies to improve quality assurance practices.</li>\n</ul>\n<p>Quality Assurance Oversight:</p>\n<ul>\n<li>Ensure compliance with regulatory requirements (GxP, ATMP) and internal quality standards throughout the product lifecycle.</li>\n<li>Act as a quality advocate between development, manufacturing, supply chain, and regulatory affairs.</li>\n<li>Provide risk-based solutions to issues that arise to ensure compliance to phase-appropriate Quality requirements, while also ensuring the flexibility to continue with process knowledge and development.</li>\n<li>Implement advanced risk management strategies, including conducting investigations (e.g., FMEA) and managing risk profiles associated with CGT products.</li>\n<li>Identify and mitigate potential risks related to quality assurance processes.</li>\n<li>Implement global CGT program elements into local procedures to ensure compliance to CGT program expectations.</li>\n<li>Analyze past performance to prevent recurrence of issues and improve future outcomes.</li>\n<li>Promote a culture of continuous learning and improvement within the quality assurance team.</li>\n<li>Engage directly in daily quality assurance activities and provide hands-on support to team members as needed.</li>\n<li>Maintain a practical mindset when addressing unforeseen problems and developing solutions.</li>\n</ul>\n<p>Regulatory Knowledge and Support:</p>\n<ul>\n<li>Ensure all documentation and processes are aligned with regulatory standards and best practices.</li>\n<li>Collaborate with Regulatory stakeholders to support the Regulatory filing and documentation needs.</li>\n</ul>\n<p>Communication and Collaboration:</p>\n<ul>\n<li>Actively participate as the Quality point of contact for projects by ensuring timely completion of tasks and deliverables.</li>\n<li>Coordinate cross-functional teams to address quality-related challenges and ensure alignment with project goals.</li>\n<li>Ensure timely communication of project milestone deliverables when at risk.</li>\n<li>Foster open communication and collaboration among team members and stakeholders, encouraging discussion and input during meetings.</li>\n<li>Communicate effectively with regulatory agencies and internal teams to ensure clarity on quality expectations.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Experience with CGT Products: Trackable experience in CGT Quality Assurance and familiarity with manufacturing processes and release criteria for CGT products.</li>\n<li>Practical Mindset in ATMP Advancement: The candidate should have a hands-on approach to advancing Advanced Therapy Medicinal Products (ATMPs) with a focus on practical solutions.</li>\n<li>Project Management Skills: Strong project management capabilities to handle both tactical and strategic elements of the role.</li>\n<li>Flexibility and Agility: The ability to adapt to changing environments and pivot strategies as necessary.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Degree in Life Sciences, Engineering, or a related field is preferred.</li>\n<li>Advanced degrees (Master&#39;s or PhD).</li>\n<li>Proven track record in Quality Assurance specifically within Cell and Gene Therapy (CGT) or Advanced Therapy Medicinal Products (ATMP).</li>\n<li>Experience with both late-stage development and marketed CGT products.</li>\n<li>Strong background in Quality Management Systems (QMS) and experience with quality audits.</li>\n<li>Knowledge of critical quality attributes and process parameters relevant to CGT products.</li>\n<li>Experience with advanced risk management techniques, including FMEA (Failure Mode and Effects Analysis).</li>\n<li>Ability to manage risk effectively throughout the product lifecycle.</li>\n<li>Experience in managing timelines and deliverables related to quality assurance.</li>\n<li>Excellent verbal and written communication skills to interact with various stakeholders, including regulatory agencies.</li>\n<li>Ability to communicate complex information clearly and effectively across different teams and cultures.</li>\n<li>Strong interpersonal skills to foster collaboration and act as a quality advocate among development, manufacturing, and regulatory teams.</li>\n<li>Ability to build relationships and work cross-functionally within an organization.</li>\n<li>Ability to adapt to changing regulatory and operational environments.</li>\n<li>Flexibility in adjusting strategies and approaches based on evolving needs and priorities.</li>\n</ul>\n<p>This is an Onsite position at our site in Berkeley, CA. Employees can expect to be paid a salary between $123,760.00 - $185,640.00. Additional compensation may include a bonus or commission (if relevant). Additional benefits include healthcare, vision, dental, retirement, PTO, sick leave, etc.</p>\n<p>This salary range is merely an estimate and may vary based on an applicant’s location, market data/ranges, skills, prior relevant experience, certain degrees and certifications, and other relevant factors.</p>\n<p>This posting will be available for application until at least 04/23/2026.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8c83efd9-b0f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976646387","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$123,760.00 - $185,640.00","x-skills-required":["CGT Quality Assurance","manufacturing processes","release criteria","regulatory requirements","quality standards","risk management","FMEA","project management","communication","collaboration","quality assurance","quality control","quality management systems","quality audits","critical quality attributes","process parameters","advanced risk management techniques"],"x-skills-preferred":["degree in Life Sciences","Engineering","related field","advanced degrees","proven track record in Quality Assurance","late-stage development","marketed CGT products","ability to manage risk","timelines and deliverables","regulatory agencies","complex information","different teams and cultures","interpersonal skills","quality advocate","building relationships","cross-functional teams","changing regulatory and operational environments","adjusting strategies"],"datePosted":"2026-04-18T22:12:10.063Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Berkeley"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"CGT Quality Assurance, manufacturing processes, release criteria, regulatory requirements, quality standards, risk management, FMEA, project management, communication, collaboration, quality assurance, quality control, quality management systems, quality audits, critical quality attributes, process parameters, advanced risk management techniques, degree in Life Sciences, Engineering, related field, advanced degrees, proven track record in Quality Assurance, late-stage development, marketed CGT products, ability to manage risk, timelines and deliverables, regulatory agencies, complex information, different teams and cultures, interpersonal skills, quality advocate, building relationships, cross-functional teams, changing regulatory and operational environments, adjusting strategies","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":123760,"maxValue":185640,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_85b01648-f2f"},"title":"Senior Director, Business Development, ECC Vehicle","description":"<p>Anduril Industries is seeking a Senior Director, Edge Compute and Communications (ECC) Business Development to lead growth and capture efforts for vehicle-integrated ECC solutions, including the Menace-X platform.</p>\n<p>In this role, you will build a cohesive, data-driven strategy to penetrate, expand, and deliver at scale our vehicle-integrated Edge Compute and Communications capabilities across all DoW services.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Developing and executing strategic initiatives to expand the adoption of Menace-X and other integrated vehicle compute solutions across DoD and allied forces.</li>\n</ul>\n<ul>\n<li>Identifying and qualifying new business opportunities where mobile ECC capabilities can address operational gaps in tactical edge environments based on operational analysis and deep customer engagement.</li>\n</ul>\n<ul>\n<li>Leading capture efforts for vehicle-based ECC programs, including support to Portfolio Acquisition Executives (PAEs), service labs, and operational units.</li>\n</ul>\n<ul>\n<li>Building and maintaining relationships with key stakeholders across the services to understand mission needs and position Anduril’s vehicle-based ECC solutions accordingly.</li>\n</ul>\n<ul>\n<li>Collaborating with engineering, product, and operations teams to shape technical roadmaps aligned with customer requirements and emerging threats.</li>\n</ul>\n<ul>\n<li>Supporting proposal development, customer briefings, and strategic planning activities.</li>\n</ul>\n<ul>\n<li>Representing Anduril’s ECC vehicle systems portfolio at industry events, demonstrations, and field exercises.</li>\n</ul>\n<ul>\n<li>Contributing to business performance tracking, forecasting, and strategic growth planning.</li>\n</ul>\n<p>Required qualifications include 7-10 years of experience in defence business development, program management, or acquisitions, strong understanding of DoD acquisition processes, government contracting, and capability transition strategies, experience with tactical vehicle platforms, mobile C2 systems, and edge compute/communications technologies, familiarity with DDIL operational environments and mobile command and control concepts, demonstrated success in capturing and growing defence programs, particularly in vehicle-based systems, excellent communication, stakeholder engagement, and cross-functional collaboration skills, ability to thrive in a fast-paced, mission-driven environment, currently possessing and able to maintain an active U.S. Secret security clearance, and ability and willingness to travel up to 50%.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_85b01648-f2f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.andurilindustries.com/","logo":"https://logos.yubhub.co/andurilindustries.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5040164007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$220,000-$292,000 USD","x-skills-required":["defence business development","program management","acquisitions","DoD acquisition processes","government contracting","capability transition strategies","tactical vehicle platforms","mobile C2 systems","edge compute/communications technologies","DDIL operational environments","mobile command and control concepts"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:09.859Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, District of Columbia, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"defence business development, program management, acquisitions, DoD acquisition processes, government contracting, capability transition strategies, tactical vehicle platforms, mobile C2 systems, edge compute/communications technologies, DDIL operational environments, mobile command and control concepts","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220000,"maxValue":292000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0d50a968-c94"},"title":"Senior Director, Business Development, ECC Vehicle","description":"<p>We are seeking a Senior Director of Business Development who is passionate about the Command, Control, Communications, and Compute (C4) missionization of ground vehicle platforms. Anduril is building the next generation of highly mobile edge infrastructure which supports the vital flow of data and processing power to the front lines, enabling intelligent, AI driven operations, even in the most contested environments.</p>\n<p>The ideal candidate has experience working with cloud and edge based systems, bringing advanced tactical communications and compute infrastructure to market, understands and can articulate the key concepts and characteristics of distributed systems, and can build a cohesive, data-driven strategy to penetrate, expand, and deliver at-scale these capabilities direct to the Department of War, its Global partners, and to other industry partners in the United States and around the world.</p>\n<p>You will create and shape the strategic long-term business opportunities based on operations analysis, user feedback generated through live force exercises, and engagements with requirements owners, capability portfolio managers, and program offices. The ideal candidate for this role also has experience working directly with ground vehicle Original Equipment Manufacturers (OEMs) to missionize platforms that are tightly integrated into the vehicle’s systems, and an understanding of how to work directly with OEMs to bring to market the fully finished and integrated systems.</p>\n<p>Key responsibilities will include:</p>\n<ul>\n<li>Developing and executing strategic initiatives to expand the adoption of Menace-X and other integrated vehicle compute solutions across DoD and allied forces.</li>\n</ul>\n<ul>\n<li>Identifying and qualifying new business opportunities where mobile ECC capabilities can address operational gaps in tactical edge environments based on operational analysis and deep customer engagement.</li>\n</ul>\n<ul>\n<li>Leading capture efforts for vehicle-based ECC programs, including support to Portfolio Acquisition Executives (PAEs), service labs, and operational units.</li>\n</ul>\n<ul>\n<li>Building and maintaining relationships with key stakeholders across the services to understand mission needs and position Anduril’s vehicle-based ECC solutions accordingly.</li>\n</ul>\n<ul>\n<li>Collaborating with engineering, product, and operations teams to shape technical roadmaps aligned with customer requirements and emerging threats.</li>\n</ul>\n<ul>\n<li>Supporting proposal development, customer briefings, and strategic planning activities.</li>\n</ul>\n<ul>\n<li>Representing Anduril’s ECC vehicle systems portfolio at industry events, demonstrations, and field exercises.</li>\n</ul>\n<ul>\n<li>Contributing to business performance tracking, forecasting, and strategic growth planning.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0d50a968-c94","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5040203007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$220,000-$292,000 USD","x-skills-required":["DoD acquisition processes","government contracting","capability transition strategies","tactical vehicle platforms","mobile C2 systems","edge compute/communications technologies","DDIL operational environments","mobile command and control concepts","capturing and growing defense programs","vehicle-based systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:48:40.764Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"DoD acquisition processes, government contracting, capability transition strategies, tactical vehicle platforms, mobile C2 systems, edge compute/communications technologies, DDIL operational environments, mobile command and control concepts, capturing and growing defense programs, vehicle-based systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220000,"maxValue":292000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1044b51e-cc6"},"title":"Senior Manager, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs. The role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land. Our Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Lead teams across autonomy, integration, and testing by aligning technical efforts, resolving cross-functional challenges, and driving mission-focused execution.</li>\n<li>Develop advanced perception algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks by integrating data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques.</li>\n<li>Develop state estimation capabilities by designing and refining algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs.</li>\n<li>Analyze and utilize sensor ICDs to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance by tuning and evaluating perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration by working closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings by leveraging synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams to ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing by contributing novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement – Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience.</li>\n<li>Typically requires a minimum of 10 years of related experience with a Bachelor’s degree; or 9 years and a Master’s degree; or 7 years with a PhD; or equivalent work experience.</li>\n<li>7+ years of experience in Unmanned Systems programs in the DoD or applied R&amp;D.</li>\n<li>2+ years of people leadership experience.</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models.</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches.</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications.</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs.</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams.</li>\n<li>Ability to obtain a SECRET clearance.</li>\n</ul>\n<p><strong>Preferences:</strong></p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems.</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks.</li>\n<li>Experience deploying perception software on SWaP-constrained platforms.</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments.</li>\n<li>Understanding of sensing challenges in denied or degraded conditions.</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1044b51e-cc6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/cebc0dd3-ffbf-4013-a2ad-ae32732cabd3","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$229,233 - $343,849 a year","x-skills-required":["BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree","10+ years of related experience","7+ years of experience in Unmanned Systems programs in the DoD or applied R&D","2+ years of people leadership experience","Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models","Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches","Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications","Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs","Proficiency with version control, debugging, and test-driven development in cross-functional teams","Ability to obtain a SECRET clearance"],"x-skills-preferred":["Hands-on integration or algorithm development with airborne sensing systems","Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks","Experience deploying perception software on SWaP-constrained platforms","Familiarity with validating perception systems during flight test events or operational environments","Understanding of sensing challenges in denied or degraded conditions","Exposure to perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:04:16.670Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, DC / San Diego, California / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, 10+ years of related experience, 7+ years of experience in Unmanned Systems programs in the DoD or applied R&D, 2+ years of people leadership experience, Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models, Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches, Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications, Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs, Proficiency with version control, debugging, and test-driven development in cross-functional teams, Ability to obtain a SECRET clearance, Hands-on integration or algorithm development with airborne sensing systems, Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks, Experience deploying perception software on SWaP-constrained platforms, Familiarity with validating perception systems during flight test events or operational environments, Understanding of sensing challenges in denied or degraded conditions, Exposure to perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":229233,"maxValue":343849,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3f0b0cce-7be"},"title":"Manager, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs.\nThe role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>We are seeking a skilled and motivated manager to lead technical teams and support direct projects integrating perception solutions for defense platforms.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land.\nOur Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Multidisciplinary Team Leadership – Lead teams across autonomy, integration, and testing by aligning technical efforts, resolving cross-functional challenges, and driving mission-focused execution.</li>\n<li>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.</li>\n<li>Develop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.</li>\n<li>Analyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement , Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience</li>\n<li>Typically requires a minimum of 7 years of related experience with a Bachelor’s degree; or 5 years and a Master’s degree; or 4 years with a PhD; or equivalent work experience</li>\n<li>5+ years of experience in Unmanned Systems programs in the DoD or applied R&amp;D</li>\n<li>2+ years of people leadership experience</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models.</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches.</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications.</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs.</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams.</li>\n<li>Ability to obtain a SECRET clearance.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems.</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks.</li>\n<li>Experience deploying perception software on SWaP-constrained platforms.</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments.</li>\n<li>Understanding of sensing challenges in denied or degraded conditions.</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3f0b0cce-7be","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/1120529c-2f7d-4b27-a29b-50976c49c433","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$220,441 - $330,661 a year","x-skills-required":["BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience","Typically requires a minimum of 7 years of related experience with a Bachelor’s degree; or 5 years and a Master’s degree; or 4 years with a PhD; or equivalent work experience","5+ years of experience in Unmanned Systems programs in the DoD or applied R&D","2+ years of people leadership experience","Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models."],"x-skills-preferred":["Hands-on integration or algorithm development with airborne sensing systems","Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks","Experience deploying perception software on SWaP-constrained platforms","Familiarity with validating perception systems during flight test events or operational environments","Understanding of sensing challenges in denied or degraded conditions"],"datePosted":"2026-04-17T13:04:04.648Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, DC"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience, Typically requires a minimum of 7 years of related experience with a Bachelor’s degree; or 5 years and a Master’s degree; or 4 years with a PhD; or equivalent work experience, 5+ years of experience in Unmanned Systems programs in the DoD or applied R&D, 2+ years of people leadership experience, Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models., Hands-on integration or algorithm development with airborne sensing systems, Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks, Experience deploying perception software on SWaP-constrained platforms, Familiarity with validating perception systems during flight test events or operational environments, Understanding of sensing challenges in denied or degraded conditions","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220441,"maxValue":330661,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_841c78ea-841"},"title":"Senior Engineer, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs.\nThe role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.\nImplement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.\nDevelop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.\nAnalyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.\nOptimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.\nSupport autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.\nValidate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.\nCollaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.\nDrive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_841c78ea-841","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/d6f1d906-5c1e-4640-87f3-3e31e1b45fa6","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 - $240,000 a year","x-skills-required":["BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience","Typically requires a minimum of 5 years of related experience with a Bachelor’s degree; or 4 years and a Master’s degree; or 2 years with a PhD; or equivalent work experience","Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models","Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches","Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications","Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs","Proficiency with version control, debugging, and test-driven development in cross-functional teams","Ability to obtain a SECRET clearance"],"x-skills-preferred":["Hands-on integration or algorithm development with airborne sensing systems","Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks","Experience deploying perception software on SWaP-constrained platforms","Familiarity with validating perception systems during flight test events or operational environments","Understanding of sensing challenges in denied or degraded conditions","Exposure to perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:03:46.950Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego, California / Washington, DC / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience, Typically requires a minimum of 5 years of related experience with a Bachelor’s degree; or 4 years and a Master’s degree; or 2 years with a PhD; or equivalent work experience, Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models, Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches, Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications, Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs, Proficiency with version control, debugging, and test-driven development in cross-functional teams, Ability to obtain a SECRET clearance, Hands-on integration or algorithm development with airborne sensing systems, Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks, Experience deploying perception software on SWaP-constrained platforms, Familiarity with validating perception systems during flight test events or operational environments, Understanding of sensing challenges in denied or degraded conditions, Exposure to perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1591a9fa-143"},"title":"Manager, Software Engineering - System Update Tools","description":"<p>The Solutions function at Shield AI is tasked with developing and deploying software applications that facilitate critical &amp; advanced operational capabilities across various systems and use cases. As a Solutions Software Engineering Manager, you will lead a team responsible for the architecture, design, and development of integrated software applications that deploy software updates and optimize in-house flight operations.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Leading and supporting the team in designing and implementing reliable systems for delivering over-the-air (OTA) software updates to internal fleet &amp; customers,</li>\n<li>Overseeing development of interactive software applications that simulate aircraft operations for training &amp; demo purposes,</li>\n<li>Guiding the development of maintenance software and operator-facing applications that help staff track, schedule, and perform maintenance activities efficiently,</li>\n<li>Supporting and reviewing enhancements to tools that collect, process, and analyze flight data, providing actionable insights that improve flight safety, efficiency, and compliance with regulatory standards,</li>\n<li>Collaborating with cross-functional teams to ensure all software solutions integrate smoothly with existing systems, maintain system integrity and performance, and reduce code duplication.</li>\n</ul>\n<p>Required qualifications include:</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Software Engineering, or a related field, or equivalent practical experience,</li>\n<li>6+ years of experience in software development, working on complex or distributed systems,</li>\n<li>1+ years of experience leading projects or managing engineers,</li>\n<li>Strong proficiency in Python and/or C++,</li>\n<li>Experience designing and building software for deployment systems, data processing, or user-facing applications,</li>\n<li>Experience collaborating with cross-functional teams such as DevOps, Integration &amp; Test, or similar.</li>\n</ul>\n<p>Preferred qualifications include:</p>\n<ul>\n<li>Experience with OTA software update systems or fleet-wide software deployment,</li>\n<li>Experience developing simulation or training applications,</li>\n<li>Experience building maintenance or operator-facing workflow tools,</li>\n<li>Experience working with flight data or similar operational data systems,</li>\n<li>Familiarity with integrating software across multiple systems and environments,</li>\n<li>Experience working in high-reliability, safety-critical, or operational environments,</li>\n<li>Experience working with or supporting testing, validation, or integration efforts.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1591a9fa-143","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/bfcf87d0-60b0-4769-bc65-4a5544b43278","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 - $240,000 a year","x-skills-required":["Python","C++","Software development","Complex systems","Project management","Team leadership","DevOps","Integration & Test"],"x-skills-preferred":["OTA software update systems","Fleet-wide software deployment","Simulation or training applications","Maintenance or operator-facing workflow tools","Flight data or similar operational data systems","Integrating software across multiple systems and environments","High-reliability, safety-critical, or operational environments"],"datePosted":"2026-04-17T13:03:46.934Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dallas"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, C++, Software development, Complex systems, Project management, Team leadership, DevOps, Integration & Test, OTA software update systems, Fleet-wide software deployment, Simulation or training applications, Maintenance or operator-facing workflow tools, Flight data or similar operational data systems, Integrating software across multiple systems and environments, High-reliability, safety-critical, or operational environments","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5f911dd8-860"},"title":"Senior Staff Engineer, Software - Perception","description":"<p>This role is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs. The role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land. Our Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.</li>\n<li>Develop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.</li>\n<li>Analyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement , Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience</li>\n<li>Typically requires a minimum of 10 years of related experience with a Bachelor’s degree; or 9 years and a Master’s degree; or 7 years with a PhD; or equivalent work experience</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams</li>\n<li>Ability to obtain a SECRET clearance</li>\n</ul>\n<p><strong>Preferences:</strong></p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks</li>\n<li>Experience deploying perception software on SWaP-constrained platforms</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments</li>\n<li>Understanding of sensing challenges in denied or degraded conditions</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5f911dd8-860","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/5cf8609e-ce9a-47e9-8956-00dae756e406","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$220,800 - $331,200 a year","x-skills-required":["algorithm development","sensor fusion","state estimation","Kalman Filters","multi-target tracking","deep learning-based detection models","probabilistic or rule-based approaches","SLAM","visual-inertial odometry","sensor-fused localization","version control","debugging","test-driven development"],"x-skills-preferred":["hands-on integration with airborne sensing systems","ML frameworks such as PyTorch or Tensorflow","perception software deployment on SWaP-constrained platforms","validating perception systems during flight test events or operational environments","sensing challenges in denied or degraded conditions","perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:03:35.432Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego, California / Washington, DC / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"algorithm development, sensor fusion, state estimation, Kalman Filters, multi-target tracking, deep learning-based detection models, probabilistic or rule-based approaches, SLAM, visual-inertial odometry, sensor-fused localization, version control, debugging, test-driven development, hands-on integration with airborne sensing systems, ML frameworks such as PyTorch or Tensorflow, perception software deployment on SWaP-constrained platforms, validating perception systems during flight test events or operational environments, sensing challenges in denied or degraded conditions, perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220800,"maxValue":331200,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bed4759c-578"},"title":"Staff Engineer, Software - Perception","description":"<p>This position is ideal for an individual who thrives on building advanced perception systems that enable autonomous aircraft to operate effectively in complex and contested environments.</p>\n<p>A successful candidate will be skilled in developing real-time object detection, sensor fusion, and state estimation algorithms using data from diverse mission sensors such as EO/IR cameras, radars, and IMUs. The role requires strong algorithmic thinking, deep familiarity with airborne sensing systems, and the ability to deliver performant software in simulation and real-world conditions.</p>\n<p>Shield AI is committed to developing cutting-edge autonomy for unmanned aircraft operating across all Department of Defense (DoD) domains, including air, sea, and land. Our Perception Engineers are instrumental in creating the situational awareness that underpins autonomy, ensuring our systems understand and respond to the operational environment with speed, precision, and resilience.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Develop advanced perception algorithms , Design and implement robust algorithms for object detection, classification, and multi-target tracking across diverse sensor modalities.</li>\n<li>Implement sensor fusion frameworks , Integrate data from vision systems, radars, and other mission sensors using probabilistic and deterministic fusion techniques to generate accurate situational awareness.</li>\n<li>Develop state estimation capabilities , Design and refine algorithms for localization and pose estimation using IMU, GPS, vision, and other onboard sensing inputs to enable stable and accurate navigation.</li>\n<li>Analyze and utilize sensor ICDs , Interpret interface control documents (ICDs) and technical specifications for aircraft-mounted sensors to ensure correct data handling, interpretation, and synchronization.</li>\n<li>Optimize perception performance , Tune and evaluate perception pipelines for performance, robustness, and real-time efficiency in both simulation and real-world environments.</li>\n<li>Support autonomy integration , Work closely with autonomy, systems, and integration teams to interface perception outputs with planning, behaviors, and decision-making modules.</li>\n<li>Validate in simulated and operational settings , Leverage synthetic data, simulation environments, and field testing to validate algorithm accuracy and mission readiness.</li>\n<li>Collaborate with hardware and sensor teams , Ensure seamless integration of perception algorithms with onboard compute platforms and diverse sensor payloads.</li>\n<li>Drive innovation in airborne sensing , Contribute novel ideas and state-of-the-art techniques to advance real-time perception capabilities for unmanned aircraft operating in complex, GPS-denied, or contested environments.</li>\n<li>Travel Requirement , Members of this team typically travel around 10-15% of the year (to different office locations, customer sites, and flight integration events).</li>\n</ul>\n<p><strong>Required Qualifications:</strong></p>\n<ul>\n<li>BS/MS in Computer Science, Electrical Engineering, Mechanical Engineering, Aerospace Engineering, and/or similar degree, or equivalent practical experience</li>\n<li>Typically requires a minimum of 7 years of related experience with a Bachelor’s degree; or 5 years and a Master’s degree; or 4 years with a PhD; or equivalent work experience</li>\n<li>Background in implementing algorithms such as Kalman Filters, multi-target tracking, or deep learning-based detection models</li>\n<li>Familiarity with fusing data from radar, EO/IR cameras, or other sensors using probabilistic or rule-based approaches</li>\n<li>Familiarity with SLAM, visual-inertial odometry, or sensor-fused localization approaches in real-time applications</li>\n<li>Ability to interpret and work with Interface Control Documents (ICDs) and hardware integration specs</li>\n<li>Proficiency with version control, debugging, and test-driven development in cross-functional teams</li>\n<li>Ability to obtain a SECRET clearance</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>Hands-on integration or algorithm development with airborne sensing systems</li>\n<li>Experience with ML frameworks such as PyTorch or Tensorflow, particularly for vision-based object detection or classification tasks</li>\n<li>Experience deploying perception software on SWaP-constrained platforms</li>\n<li>Familiarity with validating perception systems during flight test events or operational environments</li>\n<li>Understanding of sensing challenges in denied or degraded conditions</li>\n<li>Exposure to perception applications across air, maritime, and ground platforms</li>\n</ul>\n<p>$182,720 - $274,080 a year</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bed4759c-578","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Shield AI","sameAs":"https://www.shield.ai","logo":"https://logos.yubhub.co/shield.ai.png"},"x-apply-url":"https://jobs.lever.co/shieldai/8739c509-b6ea-4640-bcc1-c8b5b1de31b2","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$182,720 - $274,080 a year","x-skills-required":["real-time object detection","sensor fusion","state estimation algorithms","EO/IR cameras","radars","IMUs","Kalman Filters","multi-target tracking","deep learning-based detection models","probabilistic or rule-based approaches","SLAM","visual-inertial odometry","sensor-fused localization","Interface Control Documents","hardware integration specs","version control","debugging","test-driven development"],"x-skills-preferred":["hands-on integration or algorithm development with airborne sensing systems","ML frameworks such as PyTorch or Tensorflow","vision-based object detection or classification tasks","SWaP-constrained platforms","validating perception systems during flight test events or operational environments","sensing challenges in denied or degraded conditions","perception applications across air, maritime, and ground platforms"],"datePosted":"2026-04-17T13:02:45.901Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego, California / Washington, DC / Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"real-time object detection, sensor fusion, state estimation algorithms, EO/IR cameras, radars, IMUs, Kalman Filters, multi-target tracking, deep learning-based detection models, probabilistic or rule-based approaches, SLAM, visual-inertial odometry, sensor-fused localization, Interface Control Documents, hardware integration specs, version control, debugging, test-driven development, hands-on integration or algorithm development with airborne sensing systems, ML frameworks such as PyTorch or Tensorflow, vision-based object detection or classification tasks, SWaP-constrained platforms, validating perception systems during flight test events or operational environments, sensing challenges in denied or degraded conditions, perception applications across air, maritime, and ground platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":182720,"maxValue":274080,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3671ab35-063"},"title":"Field Service Representative – Technical Instructor","description":"<p>You will design, develop, and deliver technical training programs focused on the sustainment, maintenance, and troubleshooting of Saronic autonomous surface vessel (ASV) systems. This role blends hands-on technical instruction, field support, and performance assessment to ensure operators and maintainers are fully prepared to sustain Saronic platforms in operational environments.</p>\n<p><strong>Technical Training &amp; Curriculum Delivery</strong></p>\n<ul>\n<li>Develop and deliver training programs focused on ASV system sustainment, maintenance procedures, diagnostics, and troubleshooting.</li>\n<li>Provide instructor-led classroom and hands-on technical instruction to internal teams, contracted customers, and DoD partners.</li>\n<li>Support the development and refinement of training materials, learning objectives, assessments, and certification standards related to platform maintenance and sustainment.</li>\n<li>Conduct training needs analysis to identify sustainment-related skill gaps across operational teams.</li>\n<li>Collaborate with Mission Operations Trainers to deliver integrated operator and maintainer training programs.</li>\n</ul>\n<p><strong>Technical Procedures &amp; Sustainment Standards</strong></p>\n<ul>\n<li>Support the development and validation of maintenance procedures and sustainment best practices for Saronic ASVs.</li>\n<li>Contribute to the documentation and refinement of maintenance Standard Operating Procedures.</li>\n<li>Assist in establishing sustainment training standards and readiness benchmarks for system maintainers.</li>\n<li>Ensure training materials remain aligned with current platform configurations, software updates, and engineering changes.</li>\n</ul>\n<p><strong>Field Support &amp; Platform Integration</strong></p>\n<ul>\n<li>Provide on-site technical support during ASV system installation, integration, and commissioning activities.</li>\n<li>Assist with troubleshooting hardware, software, and system-level issues in operational environments.</li>\n<li>Support training and operational events to ensure platform reliability and readiness.</li>\n<li>Assist engineering and product teams in validating system performance during testing and operational deployments.</li>\n</ul>\n<p><strong>Sustainment Readiness &amp; Capability Development</strong></p>\n<ul>\n<li>Train operators and maintainers to independently diagnose, troubleshoot, and repair system issues.</li>\n<li>Promote disciplined maintenance practices and reinforce system safety protocols.</li>\n<li>Identify recurring technical challenges and provide coaching to improve operator and maintainer proficiency.</li>\n<li>Ensure personnel are capable of sustaining Saronic ASVs during extended operational deployments.</li>\n</ul>\n<p><strong>Stakeholder Collaboration</strong></p>\n<ul>\n<li>Coordinate with engineering, product development, and mission operations teams to ensure sustainment training reflects current system capabilities.</li>\n<li>Provide post-training feedback and field observations to engineering and product teams.</li>\n<li>Support collaboration between operational training and sustainment training programs.</li>\n<li>Assist leadership in refining sustainment readiness standards across operational teams.</li>\n</ul>\n<p><strong>Role Requirements</strong></p>\n<ul>\n<li>Demonstrated success supporting technical systems in field or operational environments.</li>\n<li>Strong organizational and time management skills with the ability to manage multiple training events and operational priorities.</li>\n<li>Excellent interpersonal and communication skills with the ability to clearly instruct technical concepts to diverse audiences.</li>\n<li>Demonstrated problem-solving ability and a proactive approach to troubleshooting complex systems.</li>\n<li>Ability to operate effectively in dynamic operational environments.</li>\n<li>Job requires up to 50% travel.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>Bachelor’s degree in engineering, robotics, computer science, maritime systems, or equivalent technical experience preferred.</li>\n<li>Experience supporting hardware/software systems in maritime, robotics, autonomous systems, or unmanned platforms.</li>\n<li>Experience troubleshooting complex mechanical, electrical, or integrated software systems.</li>\n<li>Experience delivering hands-on technical training or instructor-led instruction in operational environments.</li>\n<li>Strong analytical and problem-solving skills with the ability to diagnose system faults under operational conditions.</li>\n<li>Excellent communication and interpersonal skills with the ability to collaborate effectively across cross-functional teams.</li>\n<li>Prior experience supporting DoD or government customers is a plus.</li>\n<li>Military technical maintenance experience is a plus.</li>\n</ul>\n<p><strong>Physical Demands</strong></p>\n<ul>\n<li>Ability to certify as “fit for work” by demonstrating ability to swim and tread water.</li>\n<li>Frequently and repetitively lift, push and carry up to 50 lbs.</li>\n<li>Will be exposed to marine environments including sun, wind, rain, cold, and saltwater spray for extended periods.</li>\n<li>Ability to perform physically demanding work including stooping, bending, kneeling, crawling, and climbing for extended periods (up to 12 hours/day).</li>\n<li>Climb and maintain balance on ladders, scaffolding, or other elevated structures.</li>\n<li>Ability to stand and walk for extended periods over varied terrain.</li>\n<li>Ability to work in shipyard environments, onboard vessels under construction, and outdoor operational settings.</li>\n<li>Wear Personal Protective Equipment and follow all prescribed safety procedures.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3671ab35-063","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Saronic Technologies","sameAs":"https://www.saronic.com/","logo":"https://logos.yubhub.co/saronic.com.png"},"x-apply-url":"https://jobs.lever.co/saronic/5331a85b-d789-4a95-919a-a91a5a19a2e9","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"contract","x-salary-range":null,"x-skills-required":["Technical training","Autonomous surface vessel (ASV) systems","Maintenance procedures","Diagnostics","Troubleshooting","Hardware/software systems","Maritime, robotics, autonomous systems, or unmanned platforms","Complex mechanical, electrical, or integrated software systems","Instructor-led instruction","Operational environments"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:56:55.531Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Diego"}},"employmentType":"CONTRACTOR","occupationalCategory":"Engineering","industry":"Technology","skills":"Technical training, Autonomous surface vessel (ASV) systems, Maintenance procedures, Diagnostics, Troubleshooting, Hardware/software systems, Maritime, robotics, autonomous systems, or unmanned platforms, Complex mechanical, electrical, or integrated software systems, Instructor-led instruction, Operational environments"}]}