{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/storage"},"x-facet":{"type":"skill","slug":"storage","display":"Storage","count":100},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ebfee925-b27"},"title":"Breakfast Commis Chef","description":"<p>We&#39;re looking for a motivated and reliable Breakfast Commis Chef to join our talented kitchen team. If you&#39;re an aspiring chef with a passion for great food and a desire to grow your skills in a high-quality environment, this is a fantastic opportunity to learn from experienced chefs and be part of a well-crafted breakfast experience enjoyed by our guests.</p>\n<p>Your main responsibilities will include preparing and cooking breakfast dishes to a consistently high standard, following recipes, portioning and presentation guidelines, supporting senior chefs with daily mise en place and service, maintaining a clean and well-organised kitchen, following all food safety standards, assisting with stock rotation, storage, and basic monitoring, using kitchen equipment safely and correctly, working efficiently during busy morning service while keeping a keen eye for detail, and contributing to a positive, supportive and professional kitchen culture.</p>\n<p>In return, you&#39;ll have the chance to work within a 2 AA Rosette standard kitchen, receive ongoing training and development from an experienced and supportive team, opportunities to progress your culinary career, staff meals and uniform, a competitive salary and benefits package, free event tickets and access to our Group Performance Share scheme.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ebfee925-b27","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Goodwood Hotel","sameAs":"https://www.goodwood.com","logo":"https://logos.yubhub.co/goodwood.com.png"},"x-apply-url":"https://www.alljobspro.com/goodwood/job-details.cfm?job=312195","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":"£26,644.80 per annum","x-skills-required":["food safety level 2","basic understanding of food hygiene","kitchen equipment safety","stock rotation","storage","mise en place","presentation guidelines","recipe following","portioning"],"x-skills-preferred":["seasonal ingredients","modern cooking techniques","positive proactive attitude","desire to grow within the culinary field"],"datePosted":"2026-04-22T17:35:26.667Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Chichester"}},"employmentType":"FULL_TIME","skills":"food safety level 2, basic understanding of food hygiene, kitchen equipment safety, stock rotation, storage, mise en place, presentation guidelines, recipe following, portioning, seasonal ingredients, modern cooking techniques, positive proactive attitude, desire to grow within the culinary field","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":26644.8,"maxValue":26644.8,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6fbac7a6-777"},"title":"Battery Design Engineer","description":"<p>As a Battery Design Engineer at AVL, you will work on the development of future battery systems. You will be responsible for creating construction solutions and validation plans, analysing results from simulations and tests, and transferring these to the construction process. You will also be involved in the assembly and commissioning of batteries before testing and delivery, as well as performing tests on both modules and packs within battery testing. Additionally, you will provide technical support in customer projects and contribute to project deliveries by reviewing and releasing technical documentation. You will be responsible for leading the work on quality documentation for battery systems and components, while driving network activities within the battery area together with AVL&#39;s global network and developing local competence. You will also support business development by contributing with expertise and participating in work with offer requests and project proposals.</p>\n<p>To be successful in this role, you will have a background in battery development and a passion for creating sustainable energy solutions. You will have at least 3-5 years of experience in developing energy storage systems, with experience from SOP projects being a plus. You will have a relevant engineering degree in Mechanical, Electrical, or Mechatronics, or equivalent experience in battery system development. You will have a good understanding of high-voltage battery design and relevant safety standards and regulations. You will also have experience in battery design at the module and pack level, as well as practical experience with CAD tools such as CATIA, Creo, or similar. You will have a valid B driving license and communicate fluently in Swedish and English.</p>\n<p>As a person, you will be self-driven, structured, and responsible. You will communicate clearly and have a good ability to present technical solutions (also in English), both internally and on-site at the customer. You will build trust through your integrity and create good cooperation in teams and networks. You will be driven by solving complex problems, working analytically and quality-consciously, and have a strong desire to develop within battery design and energy storage.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6fbac7a6-777","directApply":true,"hiringOrganization":{"@type":"Organization","name":"AVL MTC Motortestcenter AB","sameAs":"https://jobs.avl.com","logo":"https://logos.yubhub.co/jobs.avl.com.png"},"x-apply-url":"https://jobs.avl.com/job/Gothenburg-Battery-Design-Engineer/1380208733/","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["battery design","energy storage systems","CAD tools","high-voltage battery design","safety standards and regulations","battery design at module and pack level"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:34:42.184Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Gothenburg"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"battery design, energy storage systems, CAD tools, high-voltage battery design, safety standards and regulations, battery design at module and pack level"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d5bcf593-890"},"title":"Werkstudent (m/w/d) für Automatisierung/Data Analysis/Machine Learning im Kontext einer techn. Großanlage","description":"<p>As a work student, you will support the continuous operation of a technical large-scale facility. Your tasks will include data analysis, machine learning, and data storage. You will work closely with our team to develop and implement new methods and techniques to improve our operations.</p>\n<p>Our team is responsible for the continuous operation of a technical large-scale facility. We are looking for a work student to support us in our daily tasks. As a work student, you will have the opportunity to gain hands-on experience in data analysis, machine learning, and data storage.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Support the continuous operation of a technical large-scale facility</li>\n<li>Data analysis and machine learning</li>\n<li>Data storage and management</li>\n<li>Development and implementation of new methods and techniques</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Study of engineering or computer science with a focus on automation, data processing, or related fields</li>\n<li>Experience with machine learning and data analysis</li>\n<li>Good knowledge of Python</li>\n<li>Excellent communication and teamwork skills</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Opportunity to gain hands-on experience in data analysis, machine learning, and data storage</li>\n<li>Collaborative and dynamic work environment</li>\n<li>Competitive salary and benefits package</li>\n</ul>\n<p>If you are interested in this opportunity, please send your application, including your resume and cover letter, to [insert contact information].</p>\n<p>We look forward to hearing from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d5bcf593-890","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dr. Ing. h.c. F. Porsche AG","sameAs":"https://jobs.porsche.com","logo":"https://logos.yubhub.co/jobs.porsche.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=20459","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"part-time","x-salary-range":null,"x-skills-required":["Python","Machine learning","Data analysis","Data storage","Automation"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:32:27.108Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Weissach"}},"employmentType":"PART_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"Python, Machine learning, Data analysis, Data storage, Automation"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_db92817e-5c5"},"title":"Auszubildender Kfz-Mechatroniker System- und Hochvolttechnik (m/w/d)","description":"<p>Here&#39;s your chance at Porsche. Porsche Centre Hamburg sells one of the world&#39;s most successful sports car brands. Our goal is to continue this success and provide our customers with optimal care.</p>\n<p>Therefore, we&#39;re looking for you to join our team as an apprentice mechatronics technician in the field of system and high-voltage technology.</p>\n<p>Your tasks:</p>\n<ul>\n<li>You will learn about the maintenance, diagnosis, and repair of vehicles, particularly in the area of vehicle electrics and electronics.</li>\n<li>You will work with the maintenance and repair of modern conventional and electric drive technologies.</li>\n<li>You will work with the installation and commissioning of chassis and driver assistance systems in the vehicle.</li>\n<li>You will learn about updating software versions on control units and working with high-voltage storage batteries in electric vehicles.</li>\n<li>You will learn about vehicle IT, smart mobility, and data transmission systems in the vehicle.</li>\n</ul>\n<p>Start date: August 2026 Duration: 3.5 years</p>\n<p>What sets you apart:</p>\n<ul>\n<li>Abitur, middle school diploma, or very good high school diploma</li>\n<li>Certificate of aptitude test for mechatronics technicians from the Hamburg Automotive Industry Association</li>\n<li>Interest in electronics, electrical, and hybrid technology</li>\n<li>Technical understanding and manual skills</li>\n<li>Responsible and reliable work ethic</li>\n<li>High team and communication skills</li>\n<li>Strong passion for automobiles</li>\n</ul>\n<p>What we offer:</p>\n<p>Benefits that make a difference - in everyday life, in the team, and beyond.</p>\n<p>Employee catering Employees receive a meal allowance that supports flexible and affordable supply during working hours.</p>\n<p>Mobility We promote sustainable mobility by subsidizing public transportation, such as discounted Germany tickets in our company account.</p>\n<p>Discounts Employees benefit from attractive corporate benefits and subsidies for EGYM and Wellpass.</p>\n<p>Salary We appreciate our employees&#39; engagement and reward it with fair and performance-based salaries according to the collective agreement. In addition to the monthly salary, we offer voluntary special payments, Christmas and holiday pay.</p>\n<p>Employee catering: Employees receive a meal allowance that supports flexible and affordable supply during working hours.</p>\n<p>Mobility: We promote sustainable mobility by subsidizing public transportation, such as discounted Germany tickets in our company account.</p>\n<p>Discounts: Employees benefit from attractive corporate benefits and subsidies for EGYM and Wellpass.</p>\n<p>Salary: We appreciate our employees&#39; engagement and reward it with fair and performance-based salaries according to the collective agreement. In addition to the monthly salary, we offer voluntary special payments, Christmas and holiday pay.</p>\n<p>About us Porsche Niederlassung Hamburg GmbH is a 100% subsidiary of Porsche Deutschland GmbH. They have three locations in Hamburg and employ over 200 people in sales, after-sales, marketing, and administration.</p>\n<p>We make the experience and fascination of Porsche more tangible. Do you also have petrol in your blood? Then join us on our continued path to success.</p>\n<p>More information about Porsche in Hamburg can be found here: https://www.porsche-in-hamburg.de/</p>\n<p>Apply now: ?ac=application&amp;jobad_id=17281</p>\n<p>Similar jobs Hamburg Auszubildender Kfz-Mechatroniker System- und Hochvolttechnik (m/w/d) Porsche Niederlassung Hamburg GmbH</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_db92817e-5c5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Porsche Niederlassung Hamburg GmbH","sameAs":"https://jobs.porsche.com","logo":"https://logos.yubhub.co/jobs.porsche.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=17281","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["mechatronics","system and high-voltage technology","vehicle electrics and electronics","conventional and electric drive technologies","chassis and driver assistance systems","software updates","high-voltage storage batteries","vehicle IT","smart mobility","data transmission systems"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:29:24.011Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hamburg"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"mechatronics, system and high-voltage technology, vehicle electrics and electronics, conventional and electric drive technologies, chassis and driver assistance systems, software updates, high-voltage storage batteries, vehicle IT, smart mobility, data transmission systems"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_326f90c8-11f"},"title":"Senior High Frequency C++ Engineer","description":"<p>The Systematic Platform Execution &amp; Exchange Data (SPEED) Team is at the core of our organisation, powering our lowest-latency solutions for systematic and high-frequency trading. We deliver the live trading and market-data platforms used by portfolio managers and risk systems, including Latency Critical Trading (LCT), DMA OMS (Client Direct), DMA market data feeds, packet capture (PCAPs), enterprise market data, and intraday data services across latency tiers from sub-100 nanoseconds to millisecond-sensitive workflows.</p>\n<p>As a Senior HFT Developer on SPEED, you will design and build core low-latency components for order entry, market data, exchange simulation, feature extraction, and strategy containers, initially focused on delivering the full set of capabilities required for trading and research infrastructure. You will collaborate closely with system architects and quantitative researchers, operate and optimise these systems in production, and have clear opportunities to grow into technical and team leadership as the effort scales.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Build low-latency infrastructure for order entry, market data, exchange simulators, feature extraction, strategy container, and other systems.</li>\n<li>Build convenience layer tools and services to facilitate trading teams onboarding at MLP.</li>\n<li>Provide level 2 support for the systems in production.</li>\n<li>Work closely with the SPEED architect, quantitative researchers, and the business to provide high ROI solutions that are aligned with both the business and the platform strategy.</li>\n<li>Opportunities for growth in terms of leadership as effort expands.</li>\n<li>Will liaise with many other MLP teams depending on project focus.</li>\n</ul>\n<p>Qualifications/Skills Required:</p>\n<ul>\n<li>5+ years with a well-regarded HFT group, delivering production-grade, low-latency systems.</li>\n<li>Demonstrated expertise in C++ and Python for production, low-latency systems.</li>\n<li>Deep familiarity with low-level Systems: OS tuning, networking stack, user-space drivers, and kernel-bypass patterns.</li>\n<li>Strong understanding of the HFT quantitative research pipeline.</li>\n<li>Experience with HPC grids (scheduling, storage, job management) for research and production workloads.</li>\n<li>Cloud experience (AWS, GCP) is a plus.</li>\n<li>Proven ability to navigate large organisations, create cross-team synergies, and influence outcomes.</li>\n<li>High accountability and ownership; able to self-manage time, set priorities, and meet deadlines.</li>\n<li>Potential to provide technical leadership and manage a small team.</li>\n</ul>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future. We pay a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_326f90c8-11f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Unknown","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954694645","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["C++","Python","low-level Systems","OS tuning","networking stack","user-space drivers","kernel-bypass patterns","HFT quantitative research pipeline","HPC grids","scheduling","storage","job management"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:18.115Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"C++, Python, low-level Systems, OS tuning, networking stack, user-space drivers, kernel-bypass patterns, HFT quantitative research pipeline, HPC grids, scheduling, storage, job management","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_07b214a4-7b2"},"title":"Associate, Quality Control Analytical","description":"<p>We are seeking an energetic and experienced individual with cell therapy quality control background to join our Quality Control (QC) team. The successful candidate will contribute to the advancement of our product pipeline by playing a crucial role in Quality Control Operations including sample management, equipment management and reagent management in the AstraZeneca Quality Control Laboratory.</p>\n<p>This position is essential for ensuring that the laboratory meets the high standards required for CGMP Quality Control Operations. The candidate will be primarily responsible for the receipt, storage, inventory, and distribution and preparation of samples, controls, reagents and standards used in analytical testing, utilizing paper and computer-based inventory systems.</p>\n<p>The candidate will also be responsible for the operation, organization and maintenance of Temperature-Controlled Storage Units and ensure that samples, reagents, standards and controls are stored in appropriate temperature conditions utilizing paper and computer-based asset systems.</p>\n<p>The candidate will support the calibration and maintenance activities required for QC equipment using computerised asset management software.</p>\n<p>The candidate will complete and document activities in accordance with CGMPs, SOPs, and protocols.</p>\n<p>Perform compendial assays and processing of samples as needed.</p>\n<p>Report deviation events to QC management.</p>\n<p>This position is essential for ensuring that the laboratory meets the high standards required for CGMP lot release and stability testing.</p>\n<p>The role also involves contributing to the writing and revision of Standard Operating Procedures (SOPs) and other CGMP documentation.</p>\n<p>The position requires a strong background in laboratory practices, general testing methodology, assay troubleshooting, and familiarity with Quality Systems.</p>\n<p>Perform other duties as assigned.</p>\n<p>Ability to follow safety procedures outlined in the Safety, Health and Environmental requirements.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_07b214a4-7b2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Quality Control","sameAs":"https://astrazeneca.eightfold.ai","logo":"https://logos.yubhub.co/astrazeneca.eightfold.ai.png"},"x-apply-url":"https://astrazeneca.eightfold.ai/careers/job/563877689883864","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$55,920.80 - $83,881.20","x-skills-required":["cell therapy quality control","sample management","equipment management","reagent management","analytical testing","Temperature-Controlled Storage Units","CGMPs","SOPs","protocols","compendial assays","processing of samples","deviation events","Standard Operating Procedures","quality systems"],"x-skills-preferred":["experience in Cell &/or Gene Therapy","Bachelor's degree in Biology or equivalent"],"datePosted":"2026-04-18T22:12:50.113Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Tarzana, California, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Research and Development","industry":"Pharmaceuticals","skills":"cell therapy quality control, sample management, equipment management, reagent management, analytical testing, Temperature-Controlled Storage Units, CGMPs, SOPs, protocols, compendial assays, processing of samples, deviation events, Standard Operating Procedures, quality systems, experience in Cell &/or Gene Therapy, Bachelor's degree in Biology or equivalent","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":55920.8,"maxValue":83881.2,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8b447835-74a"},"title":"Senior DataOps Engineer - Revenue Management (all genders)","description":"<p><strong>Your future team</strong></p>\n<p>You&#39;ll be part of our new Dynamic Pricing &amp; Revenue Management team, working alongside a Data Scientist and a Data Analyst. Together, you will work towards one core goal: helping hosts improve occupancy and earnings through a smart, dynamic, and data-driven pricing strategy.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Data Storage &amp; Querying: S3, Redshift (with decentralized data sharing), Athena, and DuckDB.</li>\n<li>ML &amp; Model Serving: MLflow, SageMaker, and deployment APIs for model lifecycle management.</li>\n<li>Cloud &amp; DevOps: Terraform, Docker, Jenkins, and AWS EKS (Kubernetes) for scalable, resilient systems.</li>\n<li>Monitoring: ELK, Grafana, Looker, OpsGenie, and in-house tools for full visibility.</li>\n<li>Ingestion: Kafka-based event systems and tools like Airbyte and Fivetran for smooth third-party integrations.</li>\n<li>Automation &amp; AI: Extensive use of AI tools like Claude, Copilot, and Codex.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>As a Data Ops Engineer – Revenue Management, you&#39;ll be the engineering backbone that enables our Data Scientists to move from experimentation to production. You bridge the gap between data science models and reliable, scalable production systems.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Support model deployment and serving: help deploy pricing and demand models into production, building and maintaining APIs and serving infrastructure.</li>\n<li>Build and operate production pipelines: ensure data flows reliably from source to model to output, with proper monitoring and alerting.</li>\n<li>Collaborate cross-functionally: work closely with Data Scientists, Analysts, and Engineering teams to turn prototypes into production-ready solutions.</li>\n<li>Own infrastructure and tooling: set up and maintain the environments, CI/CD pipelines, and infrastructure that the team depends on.</li>\n<li>Ensure operational excellence by implementing monitoring, automated testing, and observability across the team&#39;s production systems.</li>\n<li>Migrate and productionize POC: turn experimental code into robust, maintainable Python applications.</li>\n<li>Ensure data quality, consistency, and documentation across revenue management metrics and datasets.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback.</li>\n<li>Great People: Join a team of smart, motivated, and international colleagues who challenge and support each other.</li>\n<li>Technology: Work in a modern tech environment.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized.</li>\n</ul>\n<p><strong>Experience</strong></p>\n<ul>\n<li>4+ years of experience in Software Engineering, Data Engineering, DevOps, or MLOps.</li>\n<li>Strong hands-on skills in Python , you write clean, production-quality code.</li>\n<li>Experience with CI/CD, Docker, and infrastructure-as-code (e.g., Terraform).</li>\n<li>Familiarity with cloud platforms (AWS preferred) and deploying services in production.</li>\n<li>Exposure to or interest in ML model deployment (MLflow, SageMaker, or similar) is a strong plus.</li>\n<li>Desire to learn and use cutting-edge LLM tools and agents to improve your and the entire team&#39;s productivity.</li>\n<li>A proactive, hands-on mindset: you take ownership, spot problems, and drive solutions forward.</li>\n</ul>\n<p><strong>How to apply</strong></p>\n<p>If you&#39;re excited about this opportunity, please submit your application on our careers page!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8b447835-74a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2597559","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","CI/CD","Docker","Terraform","Cloud platforms (AWS preferred)","ML model deployment (MLflow, SageMaker, or similar)"],"x-skills-preferred":["AI tools like Claude, Copilot, and Codex","Data Storage & Querying (S3, Redshift, Athena, DuckDB)","ML & Model Serving (MLflow, SageMaker, deployment APIs)","Cloud & DevOps (Terraform, Docker, Jenkins, AWS EKS)","Monitoring (ELK, Grafana, Looker, OpsGenie, in-house tools)","Ingestion (Kafka-based event systems, Airbyte, Fivetran)"],"datePosted":"2026-04-18T22:09:42.352Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, CI/CD, Docker, Terraform, Cloud platforms (AWS preferred), ML model deployment (MLflow, SageMaker, or similar), AI tools like Claude, Copilot, and Codex, Data Storage & Querying (S3, Redshift, Athena, DuckDB), ML & Model Serving (MLflow, SageMaker, deployment APIs), Cloud & DevOps (Terraform, Docker, Jenkins, AWS EKS), Monitoring (ELK, Grafana, Looker, OpsGenie, in-house tools), Ingestion (Kafka-based event systems, Airbyte, Fivetran)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3ba73370-831"},"title":"Internal Audit IT Manager","description":"<p>Ready to be pushed beyond what you think you’re capable of?</p>\n<p>At Coinbase, our mission is to increase economic freedom in the world.</p>\n<p>We’re seeking a very specific candidate who is passionate about our mission and who believes in the power of crypto and blockchain technology to update the financial system.</p>\n<p>As an Internal Audit IT Manager, you will own end-to-end delivery of complex IT and security audits across our cloud infrastructure, security operations, and crypto-native systems.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Owning end-to-end delivery of IT and security audits, from risk assessment and scoping through planning, fieldwork, testing, reporting, and issue validation,covering cloud infrastructure (AWS, GCP), security operations, identity and access management, data protection, IT asset management, vendor/third-party risk, and key in-scope products and services including blockchain infrastructure, centralized and self-hosted wallets, and cold storage.</li>\n</ul>\n<ul>\n<li>Driving AI-enabled audit execution, designing and implementing data analytics, automation, and Generative AI solutions to modernize how we audit (e.g., continuous monitoring, anomaly detection, automated evidence retrieval, AI-assisted workpaper drafting),while maintaining rigorous human-in-the-loop validation to ensure accuracy and audit-quality conclusions.</li>\n</ul>\n<ul>\n<li>Executing audits aligned with the multi-year IT and security audit roadmap, coordinating coverage with co-sourced partners and cross-functional risk initiatives while ensuring alignment with Coinbase&#39;s enterprise risk profile, technology strategy, and regulatory expectations across regions (US, EMEA, APAC).</li>\n</ul>\n<ul>\n<li>Driving high-quality, risk-based findings and executive-level reporting, distilling key themes, emerging risks, and root causes into clear, concise materials for senior management and the Chief Audit Executive,ensuring findings are appropriately documented and supported by evidence.</li>\n</ul>\n<ul>\n<li>Partnering with technology and security leadership across Engineering, Security, Infrastructure, Product, and Operations to build trusted relationships, challenge control design, and advise on pragmatic, risk-based, scalable remediation while maintaining third-line independence.</li>\n</ul>\n<ul>\n<li>Driving disciplined issue management, ensuring timely, risk-based remediation by management, high-quality root cause analysis, and validation of remediation activities,escalating delays or thematic concerns to senior leadership as needed.</li>\n</ul>\n<ul>\n<li>Evaluating and developing talent, assessing candidates and helping build a high-performing, technically credible audit team.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>7+ years of experience in IT/security internal audit, technology risk, or first-line security/engineering roles with significant controls exposure.</li>\n</ul>\n<ul>\n<li>Experience working in a fast-paced, cloud-native, or engineering-driven environment where technology and security practices evolve rapidly.</li>\n</ul>\n<ul>\n<li>Hands-on audit experience with cloud platforms (AWS, GCP), including IAM policies, security configurations, logging/monitoring, and CI/CD pipelines.</li>\n</ul>\n<ul>\n<li>AI-forward mindset with demonstrated experience applying Python, SQL, or AI tools to audit or security work, building workflows rather than just prompting.</li>\n</ul>\n<ul>\n<li>Relevant professional certifications (e.g., CISA, CISSP, CIA, CISM) required; CPA or CFE a plus.</li>\n</ul>\n<ul>\n<li>Working knowledge of key frameworks such as NIST CSF, COBIT, SOC 2, and ITIL.</li>\n</ul>\n<ul>\n<li>High EQ and collaborative style.</li>\n</ul>\n<ul>\n<li>Proven ability to translate complex technical findings into clear, executive-ready narratives for both technical and non-technical audiences.</li>\n</ul>\n<ul>\n<li>Ability to manage multiple audits and initiatives across time zones (EMEA, APAC) with minimal oversight.</li>\n</ul>\n<ul>\n<li>Demonstrated leadership and team-development experience, including mentoring, coaching, and managing direct reports.</li>\n</ul>\n<ul>\n<li>Demonstrates the ability to responsibly use generative AI tools and copilots (e.g., LibreChat, Gemini, Glean) in daily workflows, continuously learn as tools evolve, and apply human-in-the-loop practices to deliver business-ready outputs and drive measurable improvements in efficiency, cost, and quality.</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience auditing or building blockchain infrastructure, crypto custody, or wallet systems (hot/cold storage).</li>\n</ul>\n<ul>\n<li>Background in a high-growth or rapidly scaling environment with complex, evolving technology stacks.</li>\n</ul>\n<ul>\n<li>Experience with GRC platforms (Workiva, Archer, AuditBoard) or building custom audit automation tooling.</li>\n</ul>\n<ul>\n<li>Familiarity with DORA, MiCA, or crypto-specific regulatory frameworks.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3ba73370-831","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Coinbase","sameAs":"https://www.coinbase.com/","logo":"https://logos.yubhub.co/coinbase.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coinbase/jobs/7755116","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,345-$195,700 USD","x-skills-required":["IT security","Cloud infrastructure","Security operations","Identity and access management","Data protection","IT asset management","Vendor/third-party risk","Blockchain infrastructure","Centralized and self-hosted wallets","Cold storage","AI-enabled audit execution","Data analytics","Automation","Generative AI","Continuous monitoring","Anomaly detection","Automated evidence retrieval","AI-assisted workpaper drafting","Cloud platforms","IAM policies","Security configurations","Logging/monitoring","CI/CD pipelines","Python","SQL","AI tools","NIST CSF","COBIT","SOC 2","ITIL","CISA","CISSP","CIA","CISM","CPA","CFE"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:31.708Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"IT security, Cloud infrastructure, Security operations, Identity and access management, Data protection, IT asset management, Vendor/third-party risk, Blockchain infrastructure, Centralized and self-hosted wallets, Cold storage, AI-enabled audit execution, Data analytics, Automation, Generative AI, Continuous monitoring, Anomaly detection, Automated evidence retrieval, AI-assisted workpaper drafting, Cloud platforms, IAM policies, Security configurations, Logging/monitoring, CI/CD pipelines, Python, SQL, AI tools, NIST CSF, COBIT, SOC 2, ITIL, CISA, CISSP, CIA, CISM, CPA, CFE","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166345,"maxValue":195700,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b05b9f90-7d3"},"title":"Data Center Engineer, Resource Efficiency – Compute Supply","description":"<p><strong>About the Role</strong></p>\n<p>As a Power &amp; Resource Efficiency Engineer, you&#39;ll sit at the intersection of IT and facilities , building the systems, models, and control loops that optimize how we allocate and consume power, cooling, and physical capacity across our TPU/GPU fleet.</p>\n<p>You&#39;ll own the technical strategy for turning raw data center capacity into reliable, efficient compute, working across power topology, workload scheduling, and real-time telemetry to push utilization as close to the physical envelope as possible while maintaining our availability commitments.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build models that forecast consumption across electrical and mechanical subsystems, informing capacity planning, energy procurement, oversubscription targets and risks, including statistical modeling of cluster utilization, workload profiles, and failure modes.</li>\n</ul>\n<ul>\n<li>Design IT/OT interfaces that bridge compute orchestration with facility controls, enabling real-time telemetry across accelerator hardware, power distribution, cooling, and schedulers.</li>\n</ul>\n<ul>\n<li>Build and operate load management systems that use power and cooling topology to enable load management and power/thermal-aware placement to maximize throughput while meeting SLOs.</li>\n</ul>\n<ul>\n<li>Partner with data center providers to drive design optimizations and hold them accountable to SLA-grade performance standards, providing technical diligence on partner architectures.</li>\n</ul>\n<p><strong>What We&#39;re Looking For</strong></p>\n<ul>\n<li>Deep knowledge of data center power distribution and cooling architectures, and how they interact with IT load profiles. Experience with reliability engineering, SLA development, and failure-mode analysis.</li>\n</ul>\n<ul>\n<li>Proficiency in statistical modeling and simulation for infrastructure capacity or power utilization.</li>\n</ul>\n<ul>\n<li>Familiarity with SCADA/BMS/EPMS, telemetry pipelines, and control systems. Experience building software that bridges IT and OT.</li>\n</ul>\n<ul>\n<li>Exposure to accelerator deployments and their power management interfaces strongly preferred.</li>\n</ul>\n<ul>\n<li>Demand response, grid interaction, or behind-the-meter generation experience is a plus.</li>\n</ul>\n<ul>\n<li>Ability to translate between infrastructure engineering, software teams, and external partners.</li>\n</ul>\n<p><strong>Required Qualifications</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Electrical Engineering, Mechanical Engineering, Power Systems, Controls Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>5+ years of experience in data center infrastructure or facility engineering.</li>\n</ul>\n<ul>\n<li>Demonstrated experience with data center power distribution and cooling system architectures.</li>\n</ul>\n<ul>\n<li>Experience building or operating software-based power management, load scheduling, or control systems.</li>\n</ul>\n<ul>\n<li>Proficiency in Python or similar languages for statistical modeling, simulation, or automation of data center infrastructure optimizations.</li>\n</ul>\n<ul>\n<li>Familiarity with SCADA, BMS, EPMS, or industrial control systems and associated protocols (Modbus, BACnet, SNMP).</li>\n</ul>\n<ul>\n<li>Track record of cross-functional collaboration across hardware, software, and facilities teams.</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Master&#39;s or PhD in Controls, Power Systems, or related discipline and 3+ years of experience in data center infrastructure or facility engineering.</li>\n</ul>\n<ul>\n<li>Experience with accelerator-class deployments and their power management interfaces.</li>\n</ul>\n<ul>\n<li>Background in control theory, dynamical systems, or cyber-physical systems design.</li>\n</ul>\n<ul>\n<li>Experience with energy storage, microgrid integration, demand response, or behind-the-meter generation.</li>\n</ul>\n<ul>\n<li>Familiarity with reliability engineering methods.</li>\n</ul>\n<ul>\n<li>Experience with SLA development, availability modeling, or service credit frameworks.</li>\n</ul>\n<ul>\n<li>Exposure to ML/optimization techniques applied to infrastructure or energy systems.</li>\n</ul>\n<p><strong>Salary</strong></p>\n<p>The annual compensation range for this role is $320,000-$405,000 USD.</p>\n<p><strong>Benefits</strong></p>\n<p>We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with our team.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b05b9f90-7d3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5159642008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["data center power distribution","cooling architectures","IT load profiles","reliability engineering","SLA development","failure-mode analysis","statistical modeling","simulation","infrastructure capacity","power utilization","SCADA/BMS/EPMS","telemetry pipelines","control systems","accelerator deployments","power management interfaces","demand response","grid interaction","behind-the-meter generation","Python","automation","data center infrastructure optimizations","SCADA","BMS","EPMS","industrial control systems","Modbus","BACnet","SNMP"],"x-skills-preferred":["accelerator-class deployments","control theory","dynamical systems","cyber-physical systems design","energy storage","microgrid integration","reliability engineering methods","availability modeling","service credit frameworks","ML/optimization techniques"],"datePosted":"2026-04-18T15:58:06.281Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-Friendly, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data center power distribution, cooling architectures, IT load profiles, reliability engineering, SLA development, failure-mode analysis, statistical modeling, simulation, infrastructure capacity, power utilization, SCADA/BMS/EPMS, telemetry pipelines, control systems, accelerator deployments, power management interfaces, demand response, grid interaction, behind-the-meter generation, Python, automation, data center infrastructure optimizations, SCADA, BMS, EPMS, industrial control systems, Modbus, BACnet, SNMP, accelerator-class deployments, control theory, dynamical systems, cyber-physical systems design, energy storage, microgrid integration, reliability engineering methods, availability modeling, service credit frameworks, ML/optimization techniques","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7dc0b69a-5b8"},"title":"Senior Engineer, Storage Control Plane","description":"<p>We&#39;re looking for a Senior Storage Engineer to play a key role in designing, building, and operating the control plane for our high-performance AI storage platform. You&#39;ll help evolve CoreWeave&#39;s storage systems by building reliable, scalable, and high-throughput solutions that power some of the largest and innovative AI workloads in the world.</p>\n<p>This role involves close collaboration with teams across infrastructure, compute, and platform to ensure our storage services scale automatically and seamlessly while maximizing performance and reliability.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Design and implement a highly scalable multi-tenant control plane that supports CoreWeave&#39;s growing AI storage and cloud infrastructure needs.</li>\n<li>Contribute to the development of exabyte-scale, S3-compatible object storage, distributed file system and integrate dedicated storage clusters into diverse customer environments.</li>\n<li>Work with technologies such as RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, and distributed filesystems to optimize storage performance and efficiency.</li>\n<li>Participate in efforts to improve the reliability, durability, and observability of our storage stack.</li>\n<li>Collaborate with operations teams to monitor, analyze, and optimize storage systems using telemetry, metrics, and dashboards to improve performance, latency, and resilience.</li>\n<li>Work cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack.</li>\n<li>Share your knowledge and mentor other engineers on best practices in building distributed, high-performance systems.</li>\n</ul>\n<p>The ideal candidate will have:</p>\n<ul>\n<li>A Bachelor&#39;s or Master&#39;s degree in Computer Science, Engineering, or a related field.</li>\n<li>6–10 years of experience working in storage systems engineering or infrastructure.</li>\n<li>Strong hands-on experience with object storage or distributed filesystems in production environments.</li>\n<li>Experience with one or more storage protocols (e.g. S3, NFS) and file systems such as Ceph, DAOS, or similar.</li>\n<li>Proficiency in a systems programming language such as Go, C, or Rust.</li>\n<li>Familiarity with storage observability tools and telemetry pipelines (e.g., ClickHouse, Prometheus, Grafana).</li>\n<li>Solid understanding of cloud-native infrastructure, Kubernetes, and scalable system architecture.</li>\n<li>Strong debugging and problem-solving skills in distributed, high-performance environments.</li>\n<li>Clear communicator, able to work collaboratively across teams and share technical insights effectively.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7dc0b69a-5b8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4611874006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$139,000 to $204,000","x-skills-required":["object storage","distributed filesystems","RDMA","GPU Direct Storage","RoCE","InfiniBand","SPDK","cloud-native infrastructure","Kubernetes","scalable system architecture"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:57.450Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"object storage, distributed filesystems, RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, cloud-native infrastructure, Kubernetes, scalable system architecture","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":139000,"maxValue":204000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d5f768d1-df6"},"title":"Full-Stack Engineer, AI Data Platform","description":"<p>Shape the Future of AI</p>\n<p>At Labelbox, we&#39;re building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we&#39;ve been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially.</p>\n<p>We&#39;re the only company offering three integrated solutions for frontier AI development:</p>\n<ul>\n<li>Enterprise Platform &amp; Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale</li>\n</ul>\n<ul>\n<li>Frontier Data Labeling Service: Specialized data labeling through Alignerr, leveraging subject matter experts for next-generation AI models</li>\n</ul>\n<ul>\n<li>Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling</li>\n</ul>\n<p>Why Join Us</p>\n<ul>\n<li>High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You&#39;ll take on expanded responsibilities quickly, with career growth directly tied to your contributions.</li>\n</ul>\n<ul>\n<li>Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence.</li>\n</ul>\n<ul>\n<li>Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution.</li>\n</ul>\n<ul>\n<li>Continuous Growth: Every role requires continuous learning and evolution. You&#39;ll be surrounded by curious minds solving complex problems at the frontier of AI.</li>\n</ul>\n<ul>\n<li>Clear Ownership: You&#39;ll know exactly what you&#39;re responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics.</li>\n</ul>\n<p>Role Overview</p>\n<p>We’re looking for a Full-Stack AI Engineer to join our team, where you’ll build the next generation of tools for developing, evaluating, and training state-of-the-art AI systems. You will own features end to end,from user-facing experiences and APIs to backend services, data models, and infrastructure.</p>\n<p>You’ll be at the heart of our applied AI efforts, with a particular focus on human-in-the-loop systems used to generate high-quality training data for Large Language Models (LLMs) and AI agents. This includes building a platform that enables us and our customers to create and evaluate data, as well as systems that leverage LLMs to assist with reviewing, scoring, and improving human submissions.</p>\n<p>Your Impact</p>\n<ul>\n<li>Own End-to-End Product Features</li>\n</ul>\n<p>Design, build, and ship complete workflows spanning frontend UI, APIs, backend services, databases, and production infrastructure.</p>\n<ul>\n<li>Enable Human-in-the-Loop AI Training</li>\n</ul>\n<p>Build systems that allow humans to efficiently create, review, and curate high-quality training and evaluation data used in AI model development.</p>\n<ul>\n<li>Support RLHF and Preference Data Workflows</li>\n</ul>\n<p>Design and implement tooling that supports RLHF-style pipelines, including task generation, human review, scoring, aggregation, and dataset versioning.</p>\n<ul>\n<li>Leverage LLMs in the Review Loop</li>\n</ul>\n<p>Build systems that use LLMs to assist human reviewers,such as automated checks, critiques, ranking suggestions, or quality signals,while maintaining human oversight.</p>\n<ul>\n<li>Advance AI Evaluation</li>\n</ul>\n<p>Design and implement evaluation frameworks and interactive tools for LLMs and AI agents across multiple data modalities (text, images, audio, video).</p>\n<ul>\n<li>Create Intuitive, Reviewer-Focused Interfaces</li>\n</ul>\n<p>Build thoughtful, efficient user interfaces (e.g., in React) optimized for high-throughput human review, quality control, and operational workflows.</p>\n<ul>\n<li>Architect Scalable Data &amp; Service Layers</li>\n</ul>\n<p>Design APIs, backend services, and data schemas that support large-scale data creation, review, and iteration with strong guarantees around correctness and traceability.</p>\n<ul>\n<li>Solve Ambiguous, Real-World Problems</li>\n</ul>\n<p>Translate loosely defined operational and research needs into practical, scalable, end-to-end systems.</p>\n<ul>\n<li>Ensure System Reliability</li>\n</ul>\n<p>Participate in on-call rotations to monitor, troubleshoot, and resolve issues across the full stack.</p>\n<ul>\n<li>Elevate the Team</li>\n</ul>\n<p>Improve engineering practices, development processes, and documentation. Share knowledge through technical writing and design discussions.</p>\n<p>What You Bring</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Data Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>2+ years of experience in a software or machine learning engineering role.</li>\n</ul>\n<ul>\n<li>A proactive, product-focused mindset and a high degree of ownership, with a passion for building solutions that empower users.</li>\n</ul>\n<ul>\n<li>Experience using frontend frameworks like React/Redux and backend systems and technologies like Python, Java, GraphQL; familiarity with NodeJS and NestJS is a plus.</li>\n</ul>\n<ul>\n<li>Knowledge of designing and managing scalable database systems, including relational databases (e.g., PostgreSQL, MySQL), NoSQL stores (e.g., MongoDB, Cassandra), and cloud-native solutions (e.g., Google Spanner, AWS DynamoDB).</li>\n</ul>\n<ul>\n<li>Familiarity with cloud infrastructure like GCP (GCS, PubSub) and containerization (Kubernetes) is a plus.</li>\n</ul>\n<ul>\n<li>Excellent communication and collaboration skills.</li>\n</ul>\n<ul>\n<li>High proficiency in leveraging AI tools for daily development (e.g., Cursor, GitHub Copilot).</li>\n</ul>\n<ul>\n<li>Comfort and enthusiasm for working in a fast-paced, agile environment where rapid problem-solving is key.</li>\n</ul>\n<p>Bonus Points</p>\n<ul>\n<li>Experience building tools for AI/ML applications, particularly for data annotation, monitoring, or agent evaluation.</li>\n</ul>\n<ul>\n<li>Familiarity with data infrastructure components such as data pipelines, streaming systems, and storage architectures (e.g., Cloud Buckets, Key-Value Stores).</li>\n</ul>\n<ul>\n<li>Previous experience with search engines (e.g., ElasticSearch).</li>\n</ul>\n<ul>\n<li>Experience in optimizing databases for performance (e.g., schema design, indexing, query tuning) and integrating them with broader data workflows.</li>\n</ul>\n<p>Engineering at Labelbox</p>\n<p>At Labelbox Engineering, we&#39;re building a comprehensive platform that powers the future of AI development. Our team combines deep technical expertise with a passion for innovation, working at the intersection of AI infrastructure, data systems, and user experience. We believe in pushing technical boundaries while maintaining high standards of code quality and system reliability. Our engineering culture emphasizes autonomous decision-making, rapid iteration, and collaborative problem-solving. We&#39;ve cultivated an environment where engineers can take ownership of significant challenges, experiment with cutting-edge technologies, and see their solutions directly impact how leading AI labs and enterprises build the next generation of AI systems.</p>\n<p>Our Technology Stack</p>\n<p>Our engineering team works with a modern tech stack designed for scalability, performance, and developer efficiency:</p>\n<ul>\n<li>Frontend: React.js with Redux, TypeScript</li>\n</ul>\n<ul>\n<li>Backend: Node.js, TypeScript, Python, some Java &amp; Kotlin</li>\n</ul>\n<ul>\n<li>APIs: GraphQL</li>\n</ul>\n<ul>\n<li>Cloud &amp; Infrastructure: Google Cloud Platform (GCP), Kubernetes</li>\n</ul>\n<ul>\n<li>Databases: MySQL, Spanner, PostgreSQL</li>\n</ul>\n<ul>\n<li>Queueing / Streaming: Kafka, PubSub</li>\n</ul>\n<p>Labelbox strives to ensure pay parity across the organization and discuss compensation transparently. The expected annual base salary range for United States-based candidates is below. This range is not inclusive of any potential equity packages or additional benefits. Exact compensation varies based on a variety of factors, including skills and competencies, experience, and geographical location.</p>\n<p>Annual base salary range $130,000-$200,000 USD</p>\n<p>Life at Labelbox</p>\n<ul>\n<li>Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland</li>\n</ul>\n<ul>\n<li>Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility</li>\n</ul>\n<ul>\n<li>Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d5f768d1-df6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Labelbox","sameAs":"https://www.labelbox.com/","logo":"https://logos.yubhub.co/labelbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/labelbox/jobs/5019254007","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$130,000-$200,000 USD","x-skills-required":["React","Redux","Node.js","TypeScript","Python","Java","GraphQL","MySQL","PostgreSQL","Spanner","Kafka","PubSub","GCP","Kubernetes","Cloud computing","Containerization","Database management","Cloud infrastructure","API design","Backend services","Data models","Infrastructure"],"x-skills-preferred":["AI tools","Cursor","GitHub Copilot","Data annotation","Monitoring","Agent evaluation","Data infrastructure","Data pipelines","Streaming systems","Storage architectures","Search engines","ElasticSearch","Database optimization","Schema design","Indexing","Query tuning"],"datePosted":"2026-04-18T15:57:55.464Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco Bay Area"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, Redux, Node.js, TypeScript, Python, Java, GraphQL, MySQL, PostgreSQL, Spanner, Kafka, PubSub, GCP, Kubernetes, Cloud computing, Containerization, Database management, Cloud infrastructure, API design, Backend services, Data models, Infrastructure, AI tools, Cursor, GitHub Copilot, Data annotation, Monitoring, Agent evaluation, Data infrastructure, Data pipelines, Streaming systems, Storage architectures, Search engines, ElasticSearch, Database optimization, Schema design, Indexing, Query tuning","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":130000,"maxValue":200000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e0058690-78c"},"title":"Senior Software Engineer, GenAI Platform","description":"<p>As a Senior Software Engineer, you will lead the development of a large-scale GenAI Platform at Reddit.</p>\n<p>The Machine Learning Platform team at Reddit is a high-impact team that owns the infrastructure that powers recommendations, content discovery, user and content quantification, while directly impacting other teams such as Growth, Ads, Feeds, and Core Machine Learning teams.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Contributing to the design, implementation, and maintenance of the LLM Gateway, focusing on features like unified API endpoints for internal/externally hosted LLM, rate/token limit management, and intelligent failover mechanisms to boost uptime and reliability.</li>\n<li>Designing and developing ML and Generative AI systems in cloud-based production environments at scale.</li>\n<li>Building and managing enterprise-grade RAG applications using embeddings, vector search, and retrieval pipelines.</li>\n<li>Implementing and operationalizing agentic AI workflows with tool use using frameworks such as LangChain and LangGraph.</li>\n<li>Driving adoption of MLOps / LLMOps practices, including CI/CD automation, versioning, testing, and lifecycle management.</li>\n<li>Establishing best practices for observability, monitoring, evaluation, and governance of GenAI pipelines in production.</li>\n</ul>\n<p>The ideal candidate will have:</p>\n<ul>\n<li>5+ years of experience in ML Engineering, AI Platform Engineering, or Cloud AI Deployment roles.</li>\n<li>Experience operating orchestration systems such as Kubernetes at scale.</li>\n<li>Deep experience with cloud-based technologies for supporting an ML platform, including tools like AWS, Google Cloud Storage, infrastructure-as-code (Terraform), and more.</li>\n<li>Proficiency with the common programming languages and frameworks of ML, such as Go, Python, etc.</li>\n<li>Excellent communication skills with the ability to articulate technical AI concepts to non-technical stakeholders.</li>\n<li>Strong focus on scalability, reliability, performance, and ease of use.</li>\n</ul>\n<p>Benefits include comprehensive healthcare benefits, income replacement programs, 401k with employer match, global benefit programs, family planning support, gender-affirming care, mental health &amp; coaching benefits, flexible vacation &amp; paid volunteer time off, and generous paid parental leave.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e0058690-78c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7753480","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$190,800-$267,100 USD","x-skills-required":["ML Engineering","AI Platform Engineering","Cloud AI Deployment","Kubernetes","AWS","Google Cloud Storage","Terraform","Go","Python","LangChain","LangGraph"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:46.916Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"ML Engineering, AI Platform Engineering, Cloud AI Deployment, Kubernetes, AWS, Google Cloud Storage, Terraform, Go, Python, LangChain, LangGraph","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":190800,"maxValue":267100,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_db7b7a31-c41"},"title":"Staff Software Engineer, Storage","description":"<p>We&#39;re seeking a Staff Software Engineer to join our BE Platform team. As a Staff Software Engineer, you will play a key role in building and evolving control and data planes, improving underlying systems, and writing software that implements critical workflows to automate and enhance the operation of our large-scale storage infrastructure.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Design, write, and deliver software to improve the availability, scalability, latency, and efficiency of Reddit&#39;s products in Go, C++, and sometimes Python.</li>\n<li>Dive deep into the codebase of supported storage systems to understand system internals.</li>\n<li>Make system-level improvements, enhancements, and implement complex code modifications.</li>\n<li>Engage actively with the open-source community to implement and upstream changes to the OSS codebase.</li>\n<li>Contribute to the design and implementation of high-performance, large-scale distributed storage systems to power various use cases at Reddit.</li>\n<li>Collaborate closely with engineering teams and stakeholders to integrate storage capabilities into broader storage infrastructure and use cases across Reddit.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>7+ years of experience building internet-scale software, preferably with a focus on machine learning storage infrastructure.</li>\n<li>Software development experience in one or more general-purpose programming languages; Golang, Python, C++, Java.</li>\n<li>Hands-on experience implementing features, optimizations, and bug fixes to distributed storage systems.</li>\n<li>Experience contributing code improvements, features, and bug fixes to open-source (OSS) projects.</li>\n<li>Prior experience with operating a large-scale critical infrastructure system with a focus on automation and workflow development is a plus, especially in a role where they were required to be on call.</li>\n<li>Excellent communication skills to collaborate with a service-oriented team and company.</li>\n</ul>\n<p>Benefits include:</p>\n<ul>\n<li>Comprehensive healthcare benefits and income replacement programs.</li>\n<li>401k match.</li>\n<li>Family planning support.</li>\n<li>Gender-affirming care.</li>\n<li>Mental health and coaching benefits.</li>\n<li>Flexible vacation and Reddit global days off.</li>\n<li>Generous paid parental leave.</li>\n<li>Paid volunteer time off.</li>\n</ul>\n<p>This job posting may span more than one career level. In addition to base salary, this job is eligible to receive equity in the form of restricted stock units, and depending on the position offered, it may also be eligible to receive a commission.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_db7b7a31-c41","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7511698","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$217,000-$303,900 USD","x-skills-required":["Go","C++","Python","Golang","Java","Distributed storage systems","Open-source community engagement","System-level improvements","High-performance storage systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:43.541Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, WA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, C++, Python, Golang, Java, Distributed storage systems, Open-source community engagement, System-level improvements, High-performance storage systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":217000,"maxValue":303900,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_95c49f85-a98"},"title":"Staff+ Software Engineer, Observability","description":"<p><strong>About the Role</strong></p>\n<p>Anthropic is seeking talented and experienced Software Engineers to join our Observability team within the Infrastructure organization. The Observability team owns the monitoring and telemetry infrastructure that every engineer and researcher at Anthropic depends on,from metrics and logging pipelines to distributed tracing, error analytics, alerting, and the dashboards and query interfaces that make it all actionable.</p>\n<p>As Anthropic scales its infrastructure across massive GPU, TPU, and Trainium clusters, the volume and complexity of operational data is growing by orders of magnitude. We’re building next-generation observability systems,high-throughput ingest pipelines, cost-efficient columnar storage, unified query layers across signals, and agentic diagnostic tools,to ensure that engineers can detect, diagnose, and resolve issues in minutes rather than hours, even as the systems they operate become exponentially more complex.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and build scalable telemetry ingest and storage pipelines for metrics, logs, traces, and error data across Anthropic’s multi-cluster infrastructure</li>\n</ul>\n<ul>\n<li>Own and evolve core observability platforms, driving migrations and architectural improvements that improve reliability, reduce cost, and scale with organisational growth</li>\n</ul>\n<ul>\n<li>Build instrumentation libraries, SDKs, and integrations that make it easy for engineering teams to emit high-quality telemetry from their services</li>\n</ul>\n<ul>\n<li>Drive alerting and SLO infrastructure that enables teams to define, monitor, and respond to reliability targets with minimal noise</li>\n</ul>\n<ul>\n<li>Reduce mean time to detection and resolution by building cross-signal correlation, unified query interfaces, and AI-assisted diagnostic tooling</li>\n</ul>\n<ul>\n<li>Partner with Research, Inference, Product, and Infrastructure teams to ensure observability solutions meet the unique needs of each organisation</li>\n</ul>\n<p><strong>You May Be a Good Fit If You</strong></p>\n<ul>\n<li>Have 10+ years of relevant industry experience building and operating large-scale observability or monitoring infrastructure</li>\n</ul>\n<ul>\n<li>Have deep experience with at least one observability signal area (metrics, logging, tracing, or error analytics) and familiarity with the others</li>\n</ul>\n<ul>\n<li>Understand high-throughput data pipelines, columnar storage engines, and the tradeoffs involved in ingesting and querying telemetry data at scale</li>\n</ul>\n<ul>\n<li>Have experience operating or building on top of observability platforms such as Prometheus, Grafana, ClickHouse, OpenTelemetry, or similar systems</li>\n</ul>\n<ul>\n<li>Have strong proficiency in at least one of Python, Rust, or Go</li>\n</ul>\n<ul>\n<li>Have excellent communication skills and enjoy partnering with internal teams to improve their operational visibility and incident response capabilities</li>\n</ul>\n<ul>\n<li>Are excited about building foundational infrastructure and are comfortable working independently on ambiguous, high-impact technical challenges</li>\n</ul>\n<p><strong>Strong Candidates May Also Have</strong></p>\n<ul>\n<li>Experience operating metrics systems at very high cardinality (hundreds of millions of active time series or more)</li>\n</ul>\n<ul>\n<li>Experience with log storage migrations or operating columnar databases (ClickHouse, BigQuery, or similar) for analytics workloads</li>\n</ul>\n<ul>\n<li>Experience with OpenTelemetry instrumentation, collector pipelines, and tail-based sampling strategies</li>\n</ul>\n<ul>\n<li>Experience building or operating alerting platforms, on-call tooling, or SLO frameworks at scale</li>\n</ul>\n<ul>\n<li>Experience with Kubernetes-native monitoring, eBPF-based observability, or continuous profiling</li>\n</ul>\n<ul>\n<li>Interest in applying AI/LLMs to operational workflows such as automated root cause analysis, anomaly detection, or intelligent alerting</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>\n</ul>\n<ul>\n<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>\n</ul>\n<ul>\n<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>\n</ul>\n<ul>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n</ul>\n<ul>\n<li>Visa sponsorship: We do sponsor visas! However, we aren’t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We’re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_95c49f85-a98","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5102440008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"£325,000-£390,000 GBP","x-skills-required":["observability","telemetry","metrics","logging","tracing","error analytics","alerting","SLO infrastructure","cross-signal correlation","unified query interfaces","AI-assisted diagnostic tooling","Python","Rust","Go","Prometheus","Grafana","ClickHouse","OpenTelemetry"],"x-skills-preferred":["high-throughput data pipelines","columnar storage engines","Kubernetes-native monitoring","eBPF-based observability","continuous profiling","AI/LLMs","automated root cause analysis","anomaly detection","intelligent alerting"],"datePosted":"2026-04-18T15:57:27.177Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"observability, telemetry, metrics, logging, tracing, error analytics, alerting, SLO infrastructure, cross-signal correlation, unified query interfaces, AI-assisted diagnostic tooling, Python, Rust, Go, Prometheus, Grafana, ClickHouse, OpenTelemetry, high-throughput data pipelines, columnar storage engines, Kubernetes-native monitoring, eBPF-based observability, continuous profiling, AI/LLMs, automated root cause analysis, anomaly detection, intelligent alerting","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":325000,"maxValue":390000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_74bd458c-995"},"title":"Engineer, Kubernetes Core Interfaces","description":"<p>The Kubernetes Core Interfaces Team at CoreWeave builds the control plane that powers our cloud infrastructure at scale. Our team ensures that the Kubernetes platforms running CoreWeave’s GPU workloads are reliable, fault-tolerant, and easy to operate, providing deep insights and smooth experiences for users and internal teams alike.</p>\n<p>As an Engineer on the Kubernetes Core Interfaces Team, you’ll design and implement scalable solutions that simplify administration and enhance the user experience of CoreWeave’s Kubernetes platforms. You will develop Helm charts, custom controllers, API endpoints, and other control plane components, while building automation, dashboards, alerts, and testing frameworks. You’ll participate in the on-call rotation and collaborate closely with your teammates in a supportive, high-performance environment that encourages curiosity, ownership, and personal growth.</p>\n<p>Some of what you’ll work on:</p>\n<ul>\n<li>Design and implement solutions for scale, fault tolerance, and operational simplicity in CoreWeave’s Kubernetes platforms.</li>\n</ul>\n<ul>\n<li>Develop Helm charts, custom controllers, CRDs, gateways, API endpoints, and other Kubernetes control plane components.</li>\n</ul>\n<ul>\n<li>Build deployment automation, monitoring dashboards, and operational insights for Kubernetes services.</li>\n</ul>\n<ul>\n<li>Participate in the team’s on-call rotation, contributing to incident response and operational excellence.</li>\n</ul>\n<ul>\n<li>Collaborate with teammates to share ideas, provide feedback, and grow together in a high-trust environment.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>3+ years of experience in software or infrastructure engineering.</li>\n</ul>\n<ul>\n<li>Experienced in developing fault-tolerant, testable software services, primarily using Go.</li>\n</ul>\n<ul>\n<li>Familiar with Kubernetes concepts and/or experienced administrating Kubernetes clusters.</li>\n</ul>\n<ul>\n<li>Comfortable working with Linux systems, shell scripting, and Linux storage/networking stacks.</li>\n</ul>\n<ul>\n<li>Collaborative, curious, and excited to contribute to a diverse, high-performance team.</li>\n</ul>\n<p>Wondering if you’re a good fit? We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren’t a 100% skill or experience match. Here are a few qualities we’ve found compatible with our team. If some of this describes you, we’d love to talk.</p>\n<ul>\n<li>You enjoy solving complex challenges at scale and improving operational workflows.</li>\n</ul>\n<ul>\n<li>You’re curious about cloud infrastructure, Kubernetes, and high-performance systems.</li>\n</ul>\n<ul>\n<li>You thrive in a collaborative environment, sharing knowledge, and learning from teammates.</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n</ul>\n<ul>\n<li>Act Like an Owner</li>\n</ul>\n<ul>\n<li>Empower Employees</li>\n</ul>\n<ul>\n<li>Deliver Best-in-Class Client Experiences</li>\n</ul>\n<ul>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and enables the development of innovative solutions to complex problems. As we get set for takeoff, the organization&#39;s growth opportunities are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!</p>\n<p>The base salary range for this role is $109,000 to $160,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location. In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n</ul>\n<ul>\n<li>100% paid for by CoreWeave</li>\n</ul>\n<ul>\n<li>Company-paid Life Insurance</li>\n</ul>\n<ul>\n<li>Voluntary supplemental life insurance</li>\n</ul>\n<ul>\n<li>Short and long-term disability insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Account</li>\n</ul>\n<ul>\n<li>Health Savings Account</li>\n</ul>\n<ul>\n<li>Tuition Reimbursement</li>\n</ul>\n<ul>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n</ul>\n<ul>\n<li>Mental Wellness Benefits through Spring Health</li>\n</ul>\n<ul>\n<li>Family-Forming support provided by Carrot</li>\n</ul>\n<ul>\n<li>Paid Parental Leave</li>\n</ul>\n<ul>\n<li>Flexible, full-service childcare support with Kinside</li>\n</ul>\n<ul>\n<li>401(k) with a generous employer match</li>\n</ul>\n<ul>\n<li>Flexible PTO</li>\n</ul>\n<ul>\n<li>Catered lunch each day in our office and data center locations</li>\n</ul>\n<ul>\n<li>A casual work environment</li>\n</ul>\n<ul>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p>California Consumer Privacy Act - California applicants only</p>\n<p>CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information. As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com.</p>\n<p>Export Control Compliance</p>\n<p>This position requires access to export controlled information. To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without a required export authorization, or (C) eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency. CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_74bd458c-995","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4656273006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$109,000 to $160,000","x-skills-required":["Go","Kubernetes","Linux","shell scripting","Linux storage/networking stacks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:22.759Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Kubernetes, Linux, shell scripting, Linux storage/networking stacks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":109000,"maxValue":160000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d9b7d5ae-6bf"},"title":"Software Engineer, Distributed Systems","description":"<p>We&#39;re growing our team of passionate creatives and builders on a mission to make design accessible to all. Our platform helps teams bring ideas to life,whether you&#39;re brainstorming, creating a prototype, translating designs into code, or iterating with AI. From idea to product, Figma empowers teams to streamline workflows, move faster, and work together in real time from anywhere in the world.</p>\n<p>As a Software Engineer on our Infrastructure team, you’ll help design, build, and operate the systems that power our real-time collaborative design tools used by millions of people worldwide. We’re scaling fast, and we’re looking for experienced distributed systems engineers across a variety of teams. Whether you’re passionate about storage, compute orchestration, developer tooling, networking, or real-time data systems, this role offers an opportunity to shape the technical foundation of one of the most beloved design platforms in the world.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and maintain scalable and reliable infrastructure systems that support product innovation and user collaboration at scale.</li>\n</ul>\n<ul>\n<li>Architect and evolve distributed systems including storage platforms, streaming infrastructure, and compute orchestration.</li>\n</ul>\n<ul>\n<li>Improve developer experience by building internal platforms, CI/CD systems, build tools, and APIs.</li>\n</ul>\n<ul>\n<li>Collaborate across product and infrastructure teams to design secure, maintainable, and performant systems.</li>\n</ul>\n<ul>\n<li>Participate in shaping platform strategy, roadmaps, and engineering best practices across the organization.</li>\n</ul>\n<ul>\n<li>Debug and resolve complex production issues that span services and layers of the stack.</li>\n</ul>\n<ul>\n<li>Mentor engineers and foster a culture of collaboration, inclusivity, and technical excellence.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>5+ years of Software Engineering experience, specifically in backend or infrastructure engineering.</li>\n</ul>\n<ul>\n<li>Deep understanding of distributed systems concepts such as sharding, replication, consistency, and eventual convergence.</li>\n</ul>\n<ul>\n<li>Experience with cloud-native environments (AWS, GCP, or Azure), infrastructure-as-code, and container orchestration.</li>\n</ul>\n<ul>\n<li>Proficiency in languages such as Go, TypeScript, Python, Rust, or Ruby.</li>\n</ul>\n<ul>\n<li>Strong system design skills and a track record of architecting resilient production systems.</li>\n</ul>\n<ul>\n<li>Excellent communication skills, with experience collaborating across teams and mentoring others.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience scaling storage platforms (e.g., Postgres, Redis, S3, DynamoDB) or operating streaming systems like Kafka.</li>\n</ul>\n<ul>\n<li>Background in traffic management, DDoS mitigation, or service mesh technologies (e.g., Envoy, Istio).</li>\n</ul>\n<ul>\n<li>A history of developing complex, real-time distributed systems at scale.</li>\n</ul>\n<ul>\n<li>A passion for building developer productivity tools, including development environments, CI/CD pipelines, and build systems.</li>\n</ul>\n<ul>\n<li>Experience with evolving large-scale, shared developer platforms to improve reliability and developer velocity.</li>\n</ul>\n<ul>\n<li>Strong problem-solving skills and a bias for action,especially when tackling high-impact, gritty challenges.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d9b7d5ae-6bf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Figma","sameAs":"https://www.figma.com/","logo":"https://logos.yubhub.co/figma.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/figma/jobs/5552549004","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$153,000-$376,000 USD","x-skills-required":["distributed systems","cloud-native environments","infrastructure-as-code","container orchestration","Go","TypeScript","Python","Rust","Ruby","system design","resilient production systems"],"x-skills-preferred":["storage platforms","streaming infrastructure","compute orchestration","developer tooling","networking","real-time data systems","traffic management","DDoS mitigation","service mesh technologies","complex distributed systems","developer productivity tools"],"datePosted":"2026-04-18T15:56:47.168Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA • New York, NY • United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"distributed systems, cloud-native environments, infrastructure-as-code, container orchestration, Go, TypeScript, Python, Rust, Ruby, system design, resilient production systems, storage platforms, streaming infrastructure, compute orchestration, developer tooling, networking, real-time data systems, traffic management, DDoS mitigation, service mesh technologies, complex distributed systems, developer productivity tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":153000,"maxValue":376000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0540dd96-198"},"title":"Senior Software Engineer - Query Engine, Database Internals - Elasticsearch","description":"<p>We&#39;re looking for a Senior Software Engineer to join the Elasticsearch - Analytical Engine team. This globally-distributed, completely remote team of senior engineers is responsible for building new analytics capabilities in Elasticsearch&#39;s latest aggregation framework based on a completely new compute engine, and accessed via our new piped query language called ES|QL.</p>\n<p>This is a senior software engineering role that covers the design and implementation of new features, enhancements to existing features, and resolving bugs.</p>\n<p>Our company is distributed by intention. We hire the best engineers we can find wherever they are, whoever they are. We collaborate across continents every day over email, GitHub, Zoom, and Slack. At our best, we write fast, scalable, and intuitive software. We believe that the best way to do that is to empower individual engineers, code review every change, decide big things by consensus, and strive for incremental improvements.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>You&#39;ll be a full-time Elasticsearch contributor, building data-intensive new features and fixing intriguing bugs, all while making the code easier to understand. You are able to research what available data structures and algorithms work best to implement a new functionality or enhancement. Sometimes you&#39;ll need to implement a data structure or algorithm in the code base. And there will be times when you&#39;ll need to get close to the operating system and hardware.</li>\n<li>You&#39;ll work with a globally distributed team of experienced engineers focused on the search and query (ES|QL) analytics capabilities of Elasticsearch. You&#39;ll get to work with the teams that build the UI to ensure a good user experience, and you&#39;ll get to work with the teams building solutions on top of these APIs</li>\n<li>You&#39;ll be an expert in several areas of Elasticsearch, and everyone will turn to you when they have a question about them. You&#39;ll improve those areas based on your questions and your instincts.</li>\n<li>You&#39;ll work with community members from all over the world on issues and pull requests, sometimes triaging them and handing them off to other experts, and sometimes handling them yourself.</li>\n<li>You&#39;ll write idiomatic modern Java -- Elasticsearch is 99.8% Java!</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>You have strong skills in core Java and are conversant in the standard library of data structures and concurrency constructs, as well as newer features like lambdas.</li>\n<li>You have experience with software systems engineering</li>\n<li>You have a strong desire to optimize and make use of the most efficient data structures and algorithms.</li>\n<li>You work with a high level of autonomy, and are able to take on projects and guide them from beginning to end. This covers both technical design and working with other engineers to develop needed components.</li>\n<li>You&#39;re comfortable developing collaboratively. Giving and receiving feedback on code, approaches, and APIs is hard! Bonus points if you&#39;ve collaborated over the internet because that&#39;s harder. Double bonus points for asynchronous collaboration over the internet. That&#39;s even harder, but we do it anyway because it&#39;s the best way we know how to build software.</li>\n<li>You&#39;ve used several data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra and have some idea how they work and why they work that way.</li>\n<li>You have excellent verbal and written communication skills. Like we said, collaborating on the internet is hard. We try to be respectful, empathetic, and trusting in all of our interactions. And we&#39;d expect that from you too.</li>\n</ul>\n<p><strong>Bonus Points</strong></p>\n<ul>\n<li>You&#39;ve built things with Elasticsearch before.</li>\n<li>You’ve worked in the search and information retrieval space. You’re familiar with the data structures and algorithms associated with information retrieval.</li>\n<li>You’ve worked on data storage technology or have experience building data analytics capabilities.</li>\n<li>You have experience designing, leading and owning cross-functional initiatives.</li>\n<li>You&#39;ve worked with open source projects and are familiar with different styles of source control workflow and continuous integration</li>\n</ul>\n<p><strong>Compensation</strong></p>\n<p>Compensation for this role is in the form of base salary. This role does not have a variable compensation component.</p>\n<p>The typical starting salary range for new hires in this role is listed below. In select locations (including Seattle WA, Los Angeles CA, the San Francisco Bay Area CA, and the New York City Metro Area), an alternate range may apply as specified below.</p>\n<p>These ranges represent the lowest to highest salary we reasonably and in good faith believe we would pay for this role at the time of this posting. We may ultimately pay more or less than the posted range, and the ranges may be modified in the future.</p>\n<p>An employee&#39;s position within the salary range will be based on several factors including, but not limited to, relevant education, qualifications, certifications, experience, skills, geographic location, performance, and business or organizational needs.</p>\n<p>Elastic believes that employees should have the opportunity to share in the value that we create together for our shareholders. Therefore, in addition to cash compensation, this role is currently eligible to participate in Elastic&#39;s stock program. Our total rewards package also includes a company-matched 401k with dollar-for-dollar matching up to 6% of eligible earnings, along with a range of other benefits offered with a holistic emphasis on employee well-being. The typical starting salary range for this role is:$133,100-$210,600 USDThe typical starting salary range for this role in the select locations listed above is:$159,900-$252,900 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0540dd96-198","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Elastic","sameAs":"https://www.elastic.co/","logo":"https://logos.yubhub.co/elastic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/elastic/jobs/7723819","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$133,100-$210,600 USD","x-skills-required":["core Java","standard library of data structures and concurrency constructs","newer features like lambdas","software systems engineering","data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:56:47.143Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"core Java, standard library of data structures and concurrency constructs, newer features like lambdas, software systems engineering, data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":133100,"maxValue":210600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_192b8eb7-029"},"title":"Staff iOS Engineer - B2C Native Apps","description":"<p>We are looking for a Staff iOS Engineer to join our B2C Native Apps team. As a member of this team, you will be responsible for designing, developing, and maintaining high-quality iOS applications.</p>\n<p>Our team is fast-paced and agile, comprising engineers, a product manager, and designer. We work closely together to deliver innovative solutions that meet the needs of our customers.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and develop high-quality iOS applications using Swift and Objective-C</li>\n<li>Collaborate with the product manager and designer to define and prioritize features</li>\n<li>Work with the engineering team to ensure seamless integration with other components</li>\n<li>Participate in code reviews and contribute to the improvement of our codebase</li>\n<li>Mentor junior engineers and help them grow in their careers</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>8+ years of professional iOS development experience</li>\n<li>Excellent communication and collaboration skills</li>\n<li>Experience building public or internal mobile APIs/SDKs and working with Swift and Objective-C</li>\n<li>Experience with UIKit, SwiftUI, programmatic Auto Layout, and iOS design patterns (MVVM, reactive programming)</li>\n<li>Experience with Unit/UI/integration/performance testing on iOS (Quick, Nimble, XCTest, XCUITest, etc.)</li>\n<li>Experience with Realm database or similar mobile NoSQL solutions</li>\n<li>End-to-end ownership of mobile applications or SDKs</li>\n<li>Experience with mobile CI/CD pipelines (GitHub Actions)</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>1+ years of experience in identity and access management (IAM) domain, particularly with Auth0 Guardian SDK or similar MFA/authentication solutions</li>\n<li>Experience with iOS security best practices, including cryptography (RSA, CommonCrypto), biometric authentication (Face ID/Touch ID), iOS Keychain, Authentication Service framework, and secure data storage</li>\n<li>Experience with reactive programming frameworks (ReactiveSwift, Combine) and migrating legacy architectures to MVVM patterns</li>\n<li>Experience with infrastructure-as-code tools (e.g., Fastlane, Swift Package Manager, Snyk, or Terraform)</li>\n</ul>\n<p>If you are a motivated and experienced iOS engineer looking to join a dynamic team, we encourage you to apply.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_192b8eb7-029","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7598837","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["iOS development","Swift","Objective-C","UIKit","SwiftUI","programmatic Auto Layout","iOS design patterns","MVVM","reactive programming","Unit/UI/integration/performance testing","Realm database","mobile NoSQL solutions","end-to-end ownership","mobile CI/CD pipelines"],"x-skills-preferred":["identity and access management","Auth0 Guardian SDK","MFA/authentication solutions","iOS security best practices","cryptography","biometric authentication","iOS Keychain","Authentication Service framework","secure data storage","reactive programming frameworks","infrastructure-as-code tools"],"datePosted":"2026-04-18T15:56:22.845Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"iOS development, Swift, Objective-C, UIKit, SwiftUI, programmatic Auto Layout, iOS design patterns, MVVM, reactive programming, Unit/UI/integration/performance testing, Realm database, mobile NoSQL solutions, end-to-end ownership, mobile CI/CD pipelines, identity and access management, Auth0 Guardian SDK, MFA/authentication solutions, iOS security best practices, cryptography, biometric authentication, iOS Keychain, Authentication Service framework, secure data storage, reactive programming frameworks, infrastructure-as-code tools"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a46b8ad4-c66"},"title":"Staff Backend Product Software Engineer, Core","description":"<p>We believe in simplifying the way people work together. As a Staff Backend Product Software Engineer, Core, you will play a pivotal role in shaping our mission of building a more enlightened way of working where everyone can unleash their creative potential without constraints.</p>\n<p>This role owns large, ambiguous problem spaces directly tied to company metrics and translates them into a clear six-month technical roadmap, leading delivery for 15–20 engineers across team boundaries. You will collaborate on ideas that shape area-level decisions, communicate effectively at all levels, and continuously integrate evolving business context into product and platform direction.</p>\n<p>You will set and model high standards for engineering and operational excellence,driving reliability, scalability, execution rigor, and long-term maintainability across foundational systems,while remaining deeply hands-on through daily coding, design authorship, critical reviews, and architectural leadership.</p>\n<p>The role influences Core Engineering roadmaps, establishes technical direction, and brings strong instincts for high-quality consumer-facing experiences. You will mentor senior talent, raising the technical bar through coaching and example, and extend your impact through org-level technical leadership.</p>\n<p>Over time, this leader grows from a cross-team technical driver into an org-level technical strategist, optimizing for the company as a whole rather than a single scope.</p>\n<p>This team will own project Optimize Activation and Engagement Flows for Team Users. Lead a multi-year plan to modernize onboarding and activation, clean up core Dropbox surfaces, and build intelligent, adaptive systems that boost team activation, invites, and long-term engagement across platforms.</p>\n<p>Opportunity for large impact on the business helping Teams users activate on the right features to drive down churn.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a46b8ad4-c66","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dropbox","sameAs":"https://www.dropbox.com/","logo":"https://logos.yubhub.co/dropbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dropbox/jobs/7421121","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$223,400-$302,200 USD (US Zone 2)","x-skills-required":["backend development","product engineering","technical leadership","engineering operations","consumer-facing products","frontend development","mobile development","search","databases","storage","concurrency","modern ML and LLM-based APIs and agentic frameworks"],"x-skills-preferred":["consumer-facing product onboarding flow on web"],"datePosted":"2026-04-18T15:56:17.112Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - US: Select locations"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend development, product engineering, technical leadership, engineering operations, consumer-facing products, frontend development, mobile development, search, databases, storage, concurrency, modern ML and LLM-based APIs and agentic frameworks, consumer-facing product onboarding flow on web","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":223400,"maxValue":302200,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4ad18574-dd4"},"title":"Software Engineer I, Backend","description":"<p>We&#39;re looking for inquisitive and well-rounded Backend engineers to join our Monetization engineering teams. As a Backend engineer, you&#39;ll play an important role in enabling the newest technologies and experiences as we scale and evolve. You will build robust frameworks &amp; features, accelerate velocity for one of the world&#39;s largest public cloud workloads, and empower both developers and Pinners alike.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Building out the backend for Pinner-facing features to power the future of inspiration on Pinterest</li>\n<li>Contributing to and leading each step of the product development process, from ideation to implementation to release</li>\n<li>Partnering with design, product, and backend teams to build end-to-end functionality</li>\n<li>Employing automated testing to build features with a high degree of technical quality</li>\n<li>Growing as an engineer by working with world-class peers on varied and high impact projects</li>\n</ul>\n<p>We&#39;re looking for candidates with 1+ years of industry backend development experience, proficiency in common backend tech stacks, and experience building &amp; operating large scale distributed systems and/or networks.</p>\n<p>As a Pinterest engineer, you&#39;ll have the opportunity to:</p>\n<ul>\n<li>Design, develop, and operate large scale, distributed systems and networks</li>\n<li>Work with Engineering customers to understand new requirements and address them in a scalable and efficient manner</li>\n<li>Actively work to improve the developer process and experience in all phases from coding to operation</li>\n</ul>\n<p>If you&#39;re passionate about building Tomas and working with a talented team, we&#39;d love to hear from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4ad18574-dd4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/6816337","x-work-arrangement":"hybrid","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":"$120,753-$211,318 USD","x-skills-required":["Python","Java","C++","Go","RESTful API","online service","storage","caching","data processing","large scale distributed systems","networks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:55:51.955Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, WA, US; Bay Area, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, C++, Go, RESTful API, online service, storage, caching, data processing, large scale distributed systems, networks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":120753,"maxValue":211318,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8f03ad2d-96f"},"title":"Software Engineer, Research Data Platform","description":"<p>We&#39;re looking for engineers who love working directly with users and who excel at building data products. The Research Data Platform team builds the tools that Anthropic&#39;s researchers use every day to manage, query, and analyze the data that goes into training and evaluating frontier models.</p>\n<p>As a Software Engineer on the Research Data Platform team, you will:</p>\n<ul>\n<li>Build and operate data pipelines that extract data from research training runs and land it in storage systems that are easy and fast to query</li>\n<li>Work closely with researchers to design and build APIs, libraries, and web interfaces that support data management, exploration, and analysis</li>\n<li>Develop dataset management, data cataloging, and provenance tooling that researchers use in their day-to-day work</li>\n<li>Embed with research teams to understand their workflows, identify high-leverage tooling opportunities, and ship solutions quickly</li>\n<li>Collaborate with adjacent teams to build on existing systems rather than reinventing them</li>\n</ul>\n<p>We do not require prior ML or AI training experience. If you enjoy working closely with technical users, learning new domains quickly, and building tools people actually want to use, you&#39;ll pick up the research context fast.</p>\n<p>Strong candidates may also have experience with large-scale ETL, columnar storage formats, and query engines (e.g., Spark, BigQuery, DuckDB, Parquet), high-volume time series data , ingestion, storage, and efficient querying, data cataloging, lineage, or metadata management systems, or ML experiment tracking or metrics platforms.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8f03ad2d-96f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5191226008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["large-scale ETL","columnar storage formats","query engines","high-volume time series data","data cataloging","lineage","metadata management systems","ML experiment tracking"],"x-skills-preferred":["Spark","BigQuery","DuckDB","Parquet"],"datePosted":"2026-04-18T15:55:38.971Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"large-scale ETL, columnar storage formats, query engines, high-volume time series data, data cataloging, lineage, metadata management systems, ML experiment tracking, Spark, BigQuery, DuckDB, Parquet","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7da67733-e98"},"title":"Systems Engineer","description":"<p>We&#39;re seeking a Systems Engineer to join LayerZero.</p>\n<p>In this role, you&#39;ll build the foundational systems that power our execution and infrastructure layers, including Zero.</p>\n<p>This is a low-level, performance-critical role focused on runtime architecture, concurrency, and scalable execution.</p>\n<p>You will make architectural decisions that directly impact how our systems execute, scale, and remain stable under load.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Architect high-performance systems and leverage AI to implement them in Rust.</li>\n</ul>\n<ul>\n<li>Build and optimise core runtime, networking, or state management components.</li>\n</ul>\n<ul>\n<li>Profile and improve performance both at low-level (CPU, memory) and cluster-scale (network, scheduling).</li>\n</ul>\n<ul>\n<li>Debug race conditions, deadlocks, and production performance issues.</li>\n</ul>\n<ul>\n<li>Contribute to architectural decisions around scheduling, synchronisation, and system scalability.</li>\n</ul>\n<ul>\n<li>Own systems end-to-end from design through production.</li>\n</ul>\n<p>About You:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, Computer Engineering, or equivalent practical experience.</li>\n</ul>\n<ul>\n<li>2+ years of experience building systems-level or performance-critical software.</li>\n</ul>\n<ul>\n<li>Strong proficiency in systems programming and leveraging AI to build software.</li>\n</ul>\n<ul>\n<li>Deep understanding of concurrency, CPU architecture, cache hierarchy, and operating systems.</li>\n</ul>\n<ul>\n<li>Experience designing or contributing to low-level storage, compute, or network infrastructure.</li>\n</ul>\n<ul>\n<li>Strong intuition for striking a pragmatic balance between complexity for performance and simplicity for robustness.</li>\n</ul>\n<ul>\n<li>Experience using low-level profiling and debugging tools in production environments.</li>\n</ul>\n<ul>\n<li>Interest in working on distributed and execution-layer systems.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7da67733-e98","directApply":true,"hiringOrganization":{"@type":"Organization","name":"LayerZero","sameAs":"https://layerzero.xyz/","logo":"https://logos.yubhub.co/layerzero.xyz.png"},"x-apply-url":"https://job-boards.greenhouse.io/layerzerolabs/jobs/5826155004","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Rust","AI","Systems programming","Concurrency","CPU architecture","Cache hierarchy","Operating systems","Low-level storage","Compute","Network infrastructure"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:55:18.270Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver, BC"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Rust, AI, Systems programming, Concurrency, CPU architecture, Cache hierarchy, Operating systems, Low-level storage, Compute, Network infrastructure"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_78d173f7-c2c"},"title":"Senior Battery Engineer, BMS","description":"<p>We are seeking an experienced Senior Battery Engineer to join our team. As a Senior Battery Engineer, you will be responsible for leading the development of energy storage systems, from concept to production. Your expertise in battery test equipment, BMS programming, and electrical circuit design will be essential in driving the success of our projects.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Drive end-to-end energy storage system (ESS) development on extremely accelerated timelines</li>\n<li>Collaborate with program teams to define actionable ESS requirements from user story/CONOPs</li>\n<li>Execute and/or deliver the following ESS development artifacts/steps:</li>\n<li>ESS requirements, battery cell selection and validation</li>\n<li>Familiarity with ESS architecture and design concepts</li>\n<li>Battery management system (BMS) requirements, BMS HW and FW/SW design, BMS SW feature plan/roadmap, BMS validation plan at bench and HIL level</li>\n<li>Cell, module, and pack level validation plan. Support bring-up, test, debug, and vehicle integration</li>\n<li>Ramp plan from development units to production</li>\n<li>Sustain fielded ESSs with root cause analysis and data processing, drive countermeasures/design updates as needed</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Bachelor’s Degree in Electrical Engineering, Mechanical Engineering or fundamental Physical Sciences with applied application focus on energy storage technologies, with 5+ years of experience designing, testing, and troubleshooting energy storage systems.</li>\n<li>Competence with battery test equipment such as cell and pack cyclers, thermal chambers, battery health testers (EIS, ACIR, hi-pot, megger, etc)</li>\n<li>One of the following: 1) Experience BMS programming cell modeling tools 2) Electrical circuit design/integrations experience.</li>\n<li>Familiarity with common electrochemical energy storage materials, technologies and boundaries of operation</li>\n<li>Knowledge of modern battery management system capabilities and requirements in HW and SW/FW</li>\n<li>Experience in battery characterization methods: mission profile validation, equivalent-circuit model characterization, cycle/calendar life assessment, safety features, etc.</li>\n<li>Understanding of UN38.3, NAVSEA S9310 and/or equivalent (UL, IEC) battery safety standards</li>\n<li>Must be a U.S. Person due to required access to U.S. export controlled information or facilities</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Graduate Degree in Electrical Engineering, Mechanical Engineering or fundamental Physical Sciences with a focus on energy storage technologies</li>\n<li>Understanding and familiarity with MIL-STD-882, MIL-STD-810, MIL-STD-704, MIL-STD-461, MIL-STD-464</li>\n<li>Demonstrated experience in delivering robust, high-performance, energy storage systems with substantial technical ownership/contributions</li>\n<li>Demonstrated experience in battery testing, modeling, controls, and BMS integration/operation</li>\n</ul>\n<p>Salary Range: $132,000 - $213,000 USD</p>\n<p>Benefits:</p>\n<ul>\n<li>Comprehensive medical, dental, and vision plans at little to no cost to you</li>\n<li>Generous time off: Highly competitive PTO plans with a holiday hiatus in December. Caregiver &amp; Wellness Leave is available to care for family members, bond with a new baby, or address your own medical needs.</li>\n<li>Family Planning &amp; Parenting Support: Coverage for fertility treatments (e.g., IVF, preservation), adoption, and gestational carriers, along with resources to support you and your partner from planning to parenting.</li>\n<li>Mental Health Resources: Access free mental health resources 24/7, including therapy and life coaching. Additional work-life services, such as legal and financial support, are also available.</li>\n<li>Professional Development: Annual reimbursement for professional development</li>\n<li>Commuter Benefits: Company-funded commuter benefits based on your region.</li>\n<li>Relocation Assistance: Available depending on role eligibility.</li>\n<li>Retirement Savings Plan: Traditional 401(k), Roth, and after-tax (mega backdoor Roth) options.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_78d173f7-c2c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/4873664007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$132,000 - $213,000 USD","x-skills-required":["Battery Test Equipment","BMS Programming","Electrical Circuit Design","Energy Storage Systems","Battery Management System","Cell Modeling Tools","Electrochemical Energy Storage Materials","Modern Battery Management System Capabilities","Battery Characterization Methods","UN38.3","NAVSEA S9310","UL","IEC"],"x-skills-preferred":["Graduate Degree in Electrical Engineering","Mechanical Engineering","Fundamental Physical Sciences","MIL-STD-882","MIL-STD-810","MIL-STD-704","MIL-STD-461","MIL-STD-464","Battery Testing","Modeling","Controls","BMS Integration/Operation"],"datePosted":"2026-04-18T15:55:13.457Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Battery Test Equipment, BMS Programming, Electrical Circuit Design, Energy Storage Systems, Battery Management System, Cell Modeling Tools, Electrochemical Energy Storage Materials, Modern Battery Management System Capabilities, Battery Characterization Methods, UN38.3, NAVSEA S9310, UL, IEC, Graduate Degree in Electrical Engineering, Mechanical Engineering, Fundamental Physical Sciences, MIL-STD-882, MIL-STD-810, MIL-STD-704, MIL-STD-461, MIL-STD-464, Battery Testing, Modeling, Controls, BMS Integration/Operation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":132000,"maxValue":213000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2e21f6d4-2c2"},"title":"Senior Software Engineer - Database Engine Internals","description":"<p>We&#39;re looking for a Senior Software Engineer to join our team in designing and implementing next-generation systems for database engine internals. You will work on query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures, and automatic physical data optimization.</p>\n<p>Our ideal candidate has a passion for database systems, storage systems, distributed systems, language design, or performance optimization, with 5+ years of experience working in a related system. A PhD in databases or distributed systems is optional.</p>\n<p>As a member of our team, you will be motivated by delivering customer value and impact, and will have the opportunity to work on a multi-year vision with incremental deliverables.</p>\n<p>The pay range for this role is $166,000-$225,000 USD, and the total compensation package may also include eligibility for annual performance bonus, equity, and benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2e21f6d4-2c2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/5048461002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$225,000 USD","x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimization"],"x-skills-preferred":["query compilation and optimization","distributed query execution and scheduling","vectorized execution engine","data security","resource management","transaction coordination","efficient storage structures","automatic physical data optimization"],"datePosted":"2026-04-18T15:54:56.759Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimization, query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures, automatic physical data optimization","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":225000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5f66d426-bea"},"title":"Principal Software Engineer, Corporate AI","description":"<p>The Principal Software Engineer is a highly skilled expert responsible for shaping and executing the organization&#39;s intelligence vision. This role integrates expertise in Artificial Intelligence (AI), Machine Learning (ML), Automation, Data Analytics and Visualization to deliver transformative customer, partner, and colleague experiences that drive revenue growth and enhance productivity.</p>\n<p>The position defines the technical direction for intelligence initiatives, leading the design, development, and deployment of robust, scalable, and secure AI solutions while fostering innovation through emerging technologies.</p>\n<p>A critical aspect of the role is providing partnership, mentorship and technical guidance, cultivating a culture of excellence and continuous learning. Through close cross-functional collaboration across teams and stakeholders, the role ensures technical efforts are strategically aligned and deliver measurable impact.</p>\n<p>Additionally, the position plays a central role in strategic problem-solving, addressing complex challenges in intelligence systems and data pipelines, and making informed architectural decisions that ensure long-term scalability and success.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Define, drive, and communicate the technical vision for intelligence, AI, and data initiatives, ensuring alignment with CIT strategy, EPD goals, and broader organisational objectives.</li>\n</ul>\n<ul>\n<li>Take a holistic view of CIT systems and architecture to ensure they are scalable, reliable, secure, and maintainable over multiple years.</li>\n</ul>\n<ul>\n<li>Lead the design, development, and deployment of high-performance AI systems, data pipelines, and intelligent services from conception through production.</li>\n</ul>\n<ul>\n<li>Make strategic architectural decisions to address complex AI, data, and platform challenges, balancing short-term delivery with long-term resilience and scalability.</li>\n</ul>\n<ul>\n<li>Identify opportunities to simplify systems, reduce operational and security risk, and improve developer productivity.</li>\n</ul>\n<ul>\n<li>Contribute directly to prototyping, proof of concepts, and implementation of technical components when needed to validate strategy, de-risk decisions, or accelerate progress.</li>\n</ul>\n<ul>\n<li>Architect, evolve, and scale AI, automation, and intelligence platforms that enable advanced analytics, personalisation, search, and intelligent decision-making.</li>\n</ul>\n<ul>\n<li>Drive innovation in intelligence models, distributed training, optimisation techniques, and data engineering to maximise performance, quality, and business impact.</li>\n</ul>\n<ul>\n<li>Enhance search and discovery capabilities using intelligent algorithms, natural language processing, and modern data systems.</li>\n</ul>\n<ul>\n<li>Evaluate, select, and integrate emerging technologies in AI, ML, and automation to maintain a competitive and forward-looking technical posture.</li>\n</ul>\n<ul>\n<li>Partner across engineering, product, design, infrastructure, and other stakeholders to ensure intelligence initiatives directly support strategic objectives.</li>\n</ul>\n<ul>\n<li>Translate technical capabilities and advancements into clear business outcomes that improve productivity, efficiency, and growth.</li>\n</ul>\n<ul>\n<li>Resolve conflicting requirements and priorities with sound technical judgment that favours long-term organisational outcomes over local optimisation.</li>\n</ul>\n<ul>\n<li>Advocate for intelligence-driven solutions across the organisation and influence company-wide technical priorities.</li>\n</ul>\n<ul>\n<li>Act as a trusted technical advisor to senior engineering leadership, with IC6 scope extending to org-wide and EPD-level strategy.</li>\n</ul>\n<ul>\n<li>Provide mentorship and technical guidance to engineers and data scientists from mid-level through senior, fostering continuous learning and technical excellence.</li>\n</ul>\n<ul>\n<li>Serve as a technical multiplier by raising the effectiveness of surrounding teams through design reviews, code reviews, architectural guidance, and pragmatic execution.</li>\n</ul>\n<ul>\n<li>Facilitate knowledge sharing across teams through documentation, design write-ups, technical discussions, and mentorship programs.</li>\n</ul>\n<ul>\n<li>Act as a voice for engineers by synthesising feedback, surfacing gaps and risks, and communicating them clearly to leadership.</li>\n</ul>\n<ul>\n<li>Contribute to multi-year technical vision and roadmap planning, anticipating future scale, complexity, and organisational needs.</li>\n</ul>\n<ul>\n<li>Identify architectural, operational, and security risks early and mobilise proactive mitigation plans across org boundaries.</li>\n</ul>\n<ul>\n<li>Partner closely with managers, product leaders, and senior engineers to ensure ambitious initiatives remain feasible, sustainable, and well-aligned.</li>\n</ul>\n<ul>\n<li>For IC6 scope, influence technical direction beyond CIT and partner directly with senior EPD leadership on company-wide strategy.</li>\n</ul>\n<ul>\n<li>Lead and support critical, high-impact initiatives by defining technical direction, clarifying requirements, gathering estimates, and ensuring delivery against milestones.</li>\n</ul>\n<ul>\n<li>Drive execution on complex projects with significant ambiguity or high cost of failure.</li>\n</ul>\n<ul>\n<li>Improve engineering effectiveness by championing best practices such as CI, automated testing, reliability reviews, and clear ownership models.</li>\n</ul>\n<ul>\n<li>Promote a bias toward action, thoughtful experimentation, and continuous learning.</li>\n</ul>\n<ul>\n<li>Model excellence in engineering craft, collaboration, accountability, and inclusive behaviour.</li>\n</ul>\n<ul>\n<li>Lead by example in living Dropbox values, including integrity, ownership, simplicity, and inclusivity.</li>\n</ul>\n<ul>\n<li>Support hiring by interviewing, calibrating candidates against a high technical bar, and representing Dropbox authentically to candidates and partners.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>12+ years of professional experience in software engineering, with depth in areas such as intelligent workflows, enterprise-scale AI adoption, automation, or data engineering.</li>\n</ul>\n<ul>\n<li>Proven track record of leading large-scale, multi-team technical initiatives from conception to production, including solving ambiguous problems, setting technical vision, and driving impact without direct authority.</li>\n</ul>\n<ul>\n<li>Strong architectural judgment and systems thinking, with the ability to balance short-term delivery with long-term sustainability, scalability, and operational excellence.</li>\n</ul>\n<ul>\n<li>Demonstrated ability to influence across teams and disciplines through technical leadership, collaboration, and sound decision-making rather than formal authority.</li>\n</ul>\n<ul>\n<li>Experience mentoring engineers and raising the technical bar of an organisation through design reviews, code reviews, and technical guidance.</li>\n</ul>\n<ul>\n<li>Exceptional written and verbal communication skills, with the ability to clearly explain complex technical concepts, translate technical strategy to diverse audiences, and influence stakeholders at all levels.</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>Strong coding ability in at least one language commonly used in AI and data systems such as Python, Java, Go, or Scala, with hands-on experience building models, data pipelines, or scalable production services.</li>\n</ul>\n<ul>\n<li>Experience operating in platform, infrastructure, or internal tooling organisations, including leading or significantly influencing org-wide technical initiatives.</li>\n</ul>\n<ul>\n<li>Proven ability to navigate ambiguity and competing priorities, drive clarity, and make sound technical and product trade-offs in partnership with product managers.</li>\n</ul>\n<ul>\n<li>Experience collaborating cross-functionally with product, design, infrastructure, and legal or privacy stakeholders to deliver AI-powered or data-intensive products responsibly.</li>\n</ul>\n<ul>\n<li>Familiarity with AI-assisted development practices in large codebases, along with experience representing engineering externally through talks, blogs, or industry events when applicable.</li>\n</ul>\n<p><strong>Compensation:</strong></p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5f66d426-bea","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dropbox","sameAs":"https://www.dropbox.com/","logo":"https://logos.yubhub.co/dropbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dropbox/jobs/7537004","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Artificial Intelligence","Machine Learning","Automation","Data Analytics","Visualization","Python","Java","Go","Scala","Cloud Storage","File-Sharing","Software Engineering","Intelligent Workflows","Enterprise-Scale AI Adoption","Data Engineering"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:49.866Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - Canada: Select locations"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Artificial Intelligence, Machine Learning, Automation, Data Analytics, Visualization, Python, Java, Go, Scala, Cloud Storage, File-Sharing, Software Engineering, Intelligent Workflows, Enterprise-Scale AI Adoption, Data Engineering"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ea5e0d36-b62"},"title":"Staff Backend Product Software Engineer, Core","description":"<p>We&#39;re seeking a Staff Backend Product Software Engineer to join our Core team. As a Staff Engineer, you will play a pivotal role in shaping our mission of building a more enlightened way of working. You will own large, ambiguous problem spaces directly tied to company metrics and translate them into a clear six-month technical roadmap. You will lead delivery for 15–20 engineers across team boundaries and collaborate on ideas that shape area-level decisions. You will communicate effectively at all levels and continuously integrate evolving business context into product and platform direction.</p>\n<p>You will set and model high standards for engineering and operational excellence, driving reliability, scalability, execution rigor, and long-term maintainability across foundational systems. You will remain deeply hands-on through daily coding, design authorship, critical reviews, and architectural leadership. You will influence Core Engineering roadmaps, establish technical direction, and bring strong instincts for high-quality consumer-facing experiences.</p>\n<p>You will mentor senior talent, raising the technical bar through coaching and example, and extend your impact through org-level technical leadership. Over time, you will grow from a cross-team technical driver into an org-level technical strategist, optimizing for the company as a whole rather than a single scope.</p>\n<p>This team will own project Optimize Activation and Engagement Flows for Team Users. You will lead a multi-year plan to modernize onboarding and activation, clean up core Dropbox surfaces, and build intelligent, adaptive systems that boost team activation, invites, and long-term engagement across platforms. Opportunity for large impact on the business helping Teams users activate on the right features to drive down churn.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ea5e0d36-b62","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dropbox","sameAs":"https://www.dropbox.com/","logo":"https://logos.yubhub.co/dropbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dropbox/jobs/7421124","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$204,900-$277,200 CAD","x-skills-required":["Computer Science","Software Engineering","Consumer User-Facing Team","Frontend","Mobile","Search","Databases","Storage","Concurrency","Modern ML and LLM-based APIs and Agentic Frameworks"],"x-skills-preferred":["Experience working on a consumer-facing product onboarding flow on web"],"datePosted":"2026-04-18T15:54:26.605Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - Canada: Select locations"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Computer Science, Software Engineering, Consumer User-Facing Team, Frontend, Mobile, Search, Databases, Storage, Concurrency, Modern ML and LLM-based APIs and Agentic Frameworks, Experience working on a consumer-facing product onboarding flow on web","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":204900,"maxValue":277200,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9be280f4-cbc"},"title":"Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for an engineer to join our small, high-impact team responsible for architecting and scaling the core infrastructure behind distributed training pipelines, multimodal data catalogs, and intelligent processing systems that operate over petabytes of data.</p>\n<p>As a software engineer on our data infrastructure team, you&#39;ll design, build, and operate scalable, fault-tolerant infrastructure for LLM Research: distributed compute, data orchestration, and storage across modalities. You&#39;ll develop high-throughput systems for data ingestion, processing, and transformation , including training data catalogs, deduplication, quality checks, and search. You&#39;ll also build systems for traceability, reproducibility, and robust quality control at every stage of the data lifecycle.</p>\n<p>You&#39;ll collaborate with research teams to unlock new features, improve data quality, and accelerate training cycles. You&#39;ll implement and maintain monitoring and alerting to support platform reliability and performance.</p>\n<p>If you&#39;re excited by distributed systems, large-scale data mining, open-source tools like Spark, Kafka, Beam, Ray, and Delta Lake, and enjoy building from the ground up, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9be280f4-cbc","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Thinking Machines Lab","sameAs":"https://thinkingmachines.ai/","logo":"https://logos.yubhub.co/thinkingmachines.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/thinkingmachines/jobs/5013919008","x-work-arrangement":"onsite","x-experience-level":"entry|mid|senior","x-job-type":"full-time","x-salary-range":"$350,000 - $475,000 USD","x-skills-required":["backend language (Python or Rust)","distributed compute frameworks (Apache Spark or Ray)","cloud infrastructure","data lake architectures","batch and streaming pipelines"],"x-skills-preferred":["Kafka","dbt","Terraform","Airflow","web crawler","deduplication","data mining","search","file formats and storage systems"],"datePosted":"2026-04-18T15:54:00.309Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend language (Python or Rust), distributed compute frameworks (Apache Spark or Ray), cloud infrastructure, data lake architectures, batch and streaming pipelines, Kafka, dbt, Terraform, Airflow, web crawler, deduplication, data mining, search, file formats and storage systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":350000,"maxValue":475000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_039ed16f-87b"},"title":"Software Engineer - Database Engine Internals","description":"<p>We are seeking a talented Software Engineer to join our team in Belgrade, Serbia. As a member of our engineering team, you will be responsible for designing and developing next-generation systems for database engine internals. Your work will focus on query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures, and automatic physical data optimization.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Drive requirements clarity and design decisions for ambiguous problems</li>\n<li>Produce technical design documents and project plans</li>\n<li>Develop new features</li>\n<li>Mentor more junior engineers</li>\n<li>Test and rollout to production, monitoring</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Passion for database systems, storage systems, distributed systems, language design, or performance optimization</li>\n<li>Comfortable working towards a multi-year vision with incremental deliverables</li>\n<li>Customer-oriented and focused on having an impact</li>\n<li>5+ years of experience working in a related system</li>\n<li>Optional: PhD in databases or distributed systems</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Comprehensive benefits and perks that meet the needs of all employees</li>\n<li>Opportunities for professional growth and development</li>\n<li>Collaborative and dynamic work environment</li>\n</ul>\n<p>At Databricks, we strive to provide a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_039ed16f-87b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8012658002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimization","query compilation and optimization","distributed query execution and scheduling","vectorized execution engine","data security","resource management","transaction coordination","efficient storage structures","automatic physical data optimization"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:53:57.962Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Belgrade, Serbia"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimization, query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures, automatic physical data optimization"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_601c2dc5-462"},"title":"Senior Software Engineer - Distributed Data Systems","description":"<p>At Databricks, we are enabling data teams to solve the world&#39;s toughest problems by building and running the world&#39;s best data and AI infrastructure platform. Our customers use deep data insights to improve their business. We are a customer-obsessed company that leaps at every opportunity to solve technical challenges.</p>\n<p>As a software engineer on the Runtime team at Databricks, you will be building the next generation distributed data storage and processing systems that can outperform specialized SQL query engines in relational query performance, yet provide the expressiveness and programming abstractions to support diverse workloads ranging from ETL to data science.</p>\n<p>Some example projects include:</p>\n<ul>\n<li>Developing the de facto open source standard framework for big data, Apache Spark.</li>\n<li>Providing reliable and high-performance services and client libraries for storing and accessing humongous amounts of data on cloud storage backends, such as AWS S3 and Azure Blob Store.</li>\n<li>Building the next generation query optimizer and execution engine that&#39;s fast, tuning-free, scalable, and robust.</li>\n</ul>\n<p>We look for candidates with a strong foundation in algorithms and data structures and their real-world use cases, experience with distributed systems, databases, and big data systems, and a BS (or higher) in Computer Science or a related technical field.</p>\n<p>The pay range for this role is $166,000-$225,000 USD, and the total compensation package may also include eligibility for annual performance bonus, equity, and benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_601c2dc5-462","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/6544325002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$225,000 USD","x-skills-required":["Java","Scala","C++","Apache Spark","Hadoop","Distributed systems","Databases","Big data systems"],"x-skills-preferred":["Algorithms","Data structures","Real-world use cases","Cloud storage backends","Query optimizer","Execution engine"],"datePosted":"2026-04-18T15:53:54.425Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Scala, C++, Apache Spark, Hadoop, Distributed systems, Databases, Big data systems, Algorithms, Data structures, Real-world use cases, Cloud storage backends, Query optimizer, Execution engine","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":225000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8363414d-965"},"title":"Senior Director of Product Marketing, Dropbox Core","description":"<p>We&#39;re hiring a Senior Director of Product Marketing to lead our Core Dropbox portfolio. This is a senior leadership role managing a team of PMMs responsible for go-to-market strategy, product positioning, lifecycle programs, and adoption strategy across web, mobile, and desktop experiences.</p>\n<p>This person will own the product marketing vision and execution for the foundational surfaces of Dropbox: content organisation, sharing, previews, and device access. You&#39;ll work closely with senior leaders in Product, Design, Growth, and Marketing to shape strategy and drive measurable business impact.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Lead Strategic Positioning and Messaging: Own and evolve the narrative for Dropbox&#39;s Core experiences. Create clear, differentiated positioning grounded in user insights and product value.</li>\n</ul>\n<ul>\n<li>Build and Manage a High-Performing Team: Recruit, lead, and coach a team of PMMs. Set priorities, provide clarity, and build a high-trust, high-impact team culture.</li>\n</ul>\n<ul>\n<li>Own Product Marketing Strategy for Core Experiences: Define and drive go-to-market strategy for Dropbox Core. Align efforts across product, growth, brand, and lifecycle marketing teams.</li>\n</ul>\n<ul>\n<li>Champion the Voice of the Customer: Lead the team in developing deep customer and market understanding. Ensure insights are actionable and influence product, messaging, and activation strategies.</li>\n</ul>\n<ul>\n<li>Drive Product-Led Growth Initiatives: Partner with Product and Growth to identify friction in key workflows, define activation strategies, and test and optimise PLG motions.</li>\n</ul>\n<ul>\n<li>Operate as a Cross-Functional Leader: Build strong partnerships across executive stakeholders in Product, Design, Marketing, and GTM. Bring strategic clarity and strong execution across initiatives.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>10+ years in product marketing or a related field, including 3+ years in a formal people management role</li>\n</ul>\n<ul>\n<li>Experience leading GTM strategy for large-scale, cross-platform products</li>\n</ul>\n<ul>\n<li>Proven success in owning and scaling product-led growth strategies</li>\n</ul>\n<ul>\n<li>Strong leadership presence with the ability to influence senior stakeholders across product and business functions</li>\n</ul>\n<ul>\n<li>Ability to coach high-performing PMMs and grow talent</li>\n</ul>\n<ul>\n<li>Excellent communication and storytelling skills for both internal and external audiences</li>\n</ul>\n<ul>\n<li>Comfort working across data, research, and customer feedback to drive decisions</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience leading PMM functions in fast-paced, matrixed environments</li>\n</ul>\n<ul>\n<li>Familiarity with productivity, collaboration, or content storage software</li>\n</ul>\n<ul>\n<li>Demonstrated ability to operate across both PLG and SLG motions</li>\n</ul>\n<ul>\n<li>Experience influencing product strategy at the executive level</li>\n</ul>\n<p>Compensation:</p>\n<p>US Zone 1: $301,800-$408,200 USD</p>\n<p>US Zone 2: $271,600-$367,400 USD</p>\n<p>US Zone 3: $241,400-$326,600 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8363414d-965","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dropbox","sameAs":"https://www.dropbox.com/","logo":"https://logos.yubhub.co/dropbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dropbox/jobs/7127856","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$301,800-$408,200 USD (US Zone 1), $271,600-$367,400 USD (US Zone 2), $241,400-$326,600 USD (US Zone 3)","x-skills-required":["product marketing","go-to-market strategy","product positioning","lifecycle programs","adoption strategy","cross-platform products","product-led growth strategies","leadership presence","coaching","communication","storytelling","data analysis","research","customer feedback"],"x-skills-preferred":["PMM functions","productivity software","collaboration software","content storage software","PLG motions","SLG motions","product strategy"],"datePosted":"2026-04-18T15:53:38.723Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - US: All locations"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Marketing","industry":"Technology","skills":"product marketing, go-to-market strategy, product positioning, lifecycle programs, adoption strategy, cross-platform products, product-led growth strategies, leadership presence, coaching, communication, storytelling, data analysis, research, customer feedback, PMM functions, productivity software, collaboration software, content storage software, PLG motions, SLG motions, product strategy","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":241400,"maxValue":408200,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_61460f7d-087"},"title":"Associate Solutions Engineer","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>The Cloudflare Associate Solution Engineering Program is a 12-month rotational experience designed to launch your career in pre-sales engineering. You&#39;ll combine technical depth, customer problem-solving, and business acumen to make Cloudflare&#39;s technology accessible and valuable for customers across Asia-Pacific.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Shadow customer calls and technical deep-dives with Enterprise and Strategic accounts</li>\n<li>Build and deliver product demonstrations tailored to customer use cases (web security, performance, serverless computing)</li>\n<li>Participate in workshops on Cloudflare technologies: Workers, Zero Trust, DNS, DDoS mitigation, WAF</li>\n<li>Collaborate with Sales, Product, and Engineering teams to solve customer technical questions</li>\n<li>Document customer requirements and translate them into solution architectures</li>\n<li>Rotate between GCR, ANZ, and ASEAN customer teams every 4 months</li>\n<li>Contribute to internal tooling, demo environments, or solution accelerators</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>Have graduated within the past 2 years (or have equivalent demonstrated technical experience through boot camps, self-study, or professional work)</li>\n<li>Can explain core networking concepts (e.g., how DNS resolution works, what happens when you visit a URL, difference between TCP/UDP)</li>\n<li>Are available to start in July 2026 and commit to 12 months including regional rotations</li>\n<li>Communicate fluently in English (written and verbal)</li>\n<li>Can manage multiple concurrent projects with competing deadlines</li>\n<li>Are authorized to work without sponsorship</li>\n</ul>\n<p>Nice to Have</p>\n<ul>\n<li>Internship or project experience in a customer-facing, consulting, or technical sales environment</li>\n<li>Proficiency in Mandarin, Cantonese, or Bahasa Indonesia (for serving regional customers)</li>\n<li>Scripting skills in Python, JavaScript, Bash, or similar</li>\n<li>Hands-on experience with web technologies: HTML/CSS/JS, HTTP APIs, or cloud platforms (AWS/GCP/Azure)</li>\n<li>Demonstrated ownership of technical projects (GitHub portfolio, conference talks, open-source contributions)</li>\n</ul>\n<p>Technologies you&#39;ll work with:</p>\n<ul>\n<li>Cloudflare&#39;s edge network</li>\n<li>Workers (serverless)</li>\n<li>Zero Trust security</li>\n<li>DNS/CDN</li>\n<li>DDoS mitigation</li>\n<li>WAF</li>\n<li>API Gateway</li>\n<li>R2 storage</li>\n<li>Stream</li>\n<li>Images</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_61460f7d-087","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7817971","x-work-arrangement":"hybrid","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Cloudflare's edge network","Workers (serverless)","Zero Trust security","DNS/CDN","DDoS mitigation","WAF","API Gateway","R2 storage","Stream","Images"],"x-skills-preferred":["Python","JavaScript","Bash","HTML/CSS/JS","HTTP APIs","cloud platforms (AWS/GCP/Azure)"],"datePosted":"2026-04-18T15:52:46.368Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Cloudflare's edge network, Workers (serverless), Zero Trust security, DNS/CDN, DDoS mitigation, WAF, API Gateway, R2 storage, Stream, Images, Python, JavaScript, Bash, HTML/CSS/JS, HTTP APIs, cloud platforms (AWS/GCP/Azure)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2702c621-d48"},"title":"Staff Backend Product Software Engineer, Commerce Platform","description":"<p>As a Staff Backend Software Engineer focused on the Commerce Platform within the Business Platform org, you will play a critical role in building and evolving the systems that power how Dropbox monetizes its products across all sales motions.</p>\n<p>You will own and influence systems spanning Salesforce CRM, CPQ, Billing, Partner Portal, and their integrations into the broader Commerce ecosystem, driving improvements in platform architecture, system simplification, and data consistency across complex, distributed workflows.</p>\n<p>You will be responsible for driving technical strategy across a highly customized, revenue-critical platform, including planning and executing modernization efforts, measuring system reliability and data integrity, and evaluating tradeoffs between short-term business needs and long-term platform health.</p>\n<p>You will operate at the intersection of engineering, Sales, and Finance, with high visibility and ownership over decisions that directly impact revenue, customer experience, and the scalability of Dropbox’s monetization systems.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and evolve scalable backend systems that power Dropbox’s Commerce Platform, including CRM, CPQ, Billing, provisioning, and related data workflows.</li>\n</ul>\n<ul>\n<li>Define and drive technical strategy for complex, highly customized systems, balancing short-term delivery with long-term platform scalability and reliability.</li>\n</ul>\n<ul>\n<li>Lead efforts to simplify legacy architectures, reduce system fragmentation, and build generalized, product-agnostic solutions.</li>\n</ul>\n<ul>\n<li>Own end-to-end reliability of revenue-critical systems, including validation, rollout safety, monitoring, and incident response.</li>\n</ul>\n<ul>\n<li>Partner cross-functionally with Engineering, Sales, Finance, and GTM teams to translate business requirements into durable technical solutions.</li>\n</ul>\n<ul>\n<li>Identify and execute on opportunities to improve data integrity, system observability, and operational efficiency across the platform.</li>\n</ul>\n<ul>\n<li>Take ownership of large, ambiguous problem spaces and drive them from definition through execution, influencing both technical and business stakeholders.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>BS, MS, or PhD in Computer Science or a related technical field involving programming (e.g., Computer Engineering), or equivalent practical experience.</li>\n</ul>\n<ul>\n<li>12+ years of professional software development experience.</li>\n</ul>\n<ul>\n<li>Proven expertise in building and maintaining large-scale, distributed backend systems.</li>\n</ul>\n<ul>\n<li>Strong system design skills, particularly in complex, stateful, or workflow-driven systems</li>\n</ul>\n<ul>\n<li>Experience building reliable systems with a focus on validation, rollout safety, monitoring, and maintainability</li>\n</ul>\n<ul>\n<li>Experience working on systems in a monetary domain such as ecommerce, subscriptions, billing, payments, or financial systems</li>\n</ul>\n<ul>\n<li>Strong communication and collaboration skills, with experience influencing cross-functional stakeholders</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with CRM, CPQ, Billing, ecommerce, or similar enterprise workflow platforms (e.g., Salesforce)</li>\n</ul>\n<ul>\n<li>Track record of simplifying or modernizing legacy systems into scalable, maintainable architectures</li>\n</ul>\n<ul>\n<li>Experience working on revenue-critical or monetization platforms at scale</li>\n</ul>\n<ul>\n<li>Experience improving data consistency and integrity across distributed systems</li>\n</ul>\n<ul>\n<li>Demonstrated ability to influence technical direction across teams without direct authority</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2702c621-d48","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dropbox","sameAs":"https://www.dropbox.com/","logo":"https://logos.yubhub.co/dropbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dropbox/jobs/7759728","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$223,400-$302,200 USD Zone 2","x-skills-required":["Computer Science","Programming","System Design","Backend Systems","Cloud Storage","File Sharing","Salesforce","CRM","CPQ","Billing","Ecommerce","Subscriptions","Payments","Financial Systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:44.865Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - US: Select locations"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Computer Science, Programming, System Design, Backend Systems, Cloud Storage, File Sharing, Salesforce, CRM, CPQ, Billing, Ecommerce, Subscriptions, Payments, Financial Systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":223400,"maxValue":302200,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1c7dc0cb-87c"},"title":"Solutions Architect - Storage","description":"<p>As a Solutions Architect at CoreWeave, you will play a vital and dynamic role in helping customers succeed with our cloud infrastructure offerings. You will serve as the primary technical point of contact for customers, establishing strong technical relationships and ensuring their success with CoreWeave&#39;s cloud infrastructure offerings, focusing on storage technologies within high-performance compute (HPC) environments.</p>\n<p>Collaborate closely with customers to understand their unique business needs and create, prototype, and deploy tailored solutions that align with their requirements. Lead proof of concept initiatives to showcase the value and viability of CoreWeave&#39;s solutions within specific environments.</p>\n<p>Drive technical leadership and direction during customer meetings, presentations, and workshops, addressing any technical queries or concerns that arise. Act as a virtual member of CoreWeave&#39;s Storage product and engineering teams, identifying opportunities for product enhancement and collaborating with engineers to implement your suggestions.</p>\n<p>Offer valuable insights on product features, functionality, and performance, contributing regularly to discussions about product strategy and architecture. Conduct periodic technical reviews and assessments of customer workloads, pinpointing opportunities for workload optimization and suggesting suitable solutions.</p>\n<p>Stay informed of the latest developments and trends in Kubernetes, cloud computing and infrastructure, sharing your thought leadership with customers and internal stakeholders. Lead the prototyping and initiation of research and development efforts for emerging products and solutions, delivering prototypes and key insights for internal consumption.</p>\n<p>Represent CoreWeave at conferences and industry events, with occasional travel as required.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1c7dc0cb-87c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4568531006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$165,000 to $220,000","x-skills-required":["cloud computing concepts","architecture","technologies","storage solutions","Kubernetes","cloud infrastructure","high-performance compute (HPC)","storage technologies","file system protocols","infrastructure systems"],"x-skills-preferred":["code contributions to open-source inference frameworks","scripting and automation related to storage technologies","building solutions across multi-cloud environments","client or customer-facing publications/talks on latency, optimization, or advanced model-server architectures"],"datePosted":"2026-04-18T15:52:39.508Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"cloud computing concepts, architecture, technologies, storage solutions, Kubernetes, cloud infrastructure, high-performance compute (HPC), storage technologies, file system protocols, infrastructure systems, code contributions to open-source inference frameworks, scripting and automation related to storage technologies, building solutions across multi-cloud environments, client or customer-facing publications/talks on latency, optimization, or advanced model-server architectures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_92f42798-2b7"},"title":"Engineering Manager II, Cache Infra","description":"<p>About Pinterest:</p>\n<p>Millions of people around the world come to our platform to find creative ideas, dream about new possibilities and plan for memories that will last a lifetime.</p>\n<p>At Pinterest, we&#39;re on a mission to bring everyone the inspiration to create a life they love, and that starts with the people behind the product.</p>\n<p>We&#39;re looking for an experienced Engineering Manager II to lead our Cache Infra team in building and operating large-scale distributed systems.</p>\n<p>This role provides a unique opportunity to gain hands-on experience in building and operating large-scale distributed systems.</p>\n<p>The team works on challenging, exciting, and impactful projects, including new cache stack, compute modernization initiatives, and innovations related to AI.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Lead a high-performing team to build reliable, performant and efficient caching systems that operate at a huge scale and empower numerous business critical applications across the entire company.</li>\n</ul>\n<ul>\n<li>Lead a cross-functional strategic initiative to modernize the cache infra tech stack by adopting cutting-edge technologies.</li>\n</ul>\n<ul>\n<li>Maintain a high engineering standard for cache infrastructure, focusing on continuous improvement in reliability, scalability, performance, cost efficiency, and developer velocity.</li>\n</ul>\n<ul>\n<li>Drive the strategic migration of memcached from the current legacy deployment system to Kubernetes.</li>\n</ul>\n<ul>\n<li>Drive the team vision and strategy for the next 3 years and beyond.</li>\n</ul>\n<ul>\n<li>Help create a diverse and inclusive team culture that makes all feel welcome while encouraging critical thinking and candid feedback.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>1+ years of Experience leading or managing engineering teams, including setting direction, driving execution, and supporting team members&#39; growth.</li>\n</ul>\n<ul>\n<li>5+ years of relevant industry experience</li>\n</ul>\n<ul>\n<li>Strong technical background in distributed systems and storage/caching systems.</li>\n</ul>\n<ul>\n<li>Hands on experience building and operating highly available, reliable, production grade systems at scale.</li>\n</ul>\n<ul>\n<li>Experience coding in one of the following languages: Java, Python and/or C/C++.</li>\n</ul>\n<ul>\n<li>Experience recruiting and managing technical teams, including performance management.</li>\n</ul>\n<ul>\n<li>Demonstrated experience leveraging AI to accelerate development, enhance operations, and improve customer support.</li>\n</ul>\n<ul>\n<li>Exceptional collaboration skills with cross-functional partners, with the ability to navigate ambiguity, make tradeoffs, and keep stakeholders aligned on priorities and progress.</li>\n</ul>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, a related technical field, or equivalent experience.</li>\n</ul>\n<p>Relocation Statement:</p>\n<p>This position is not eligible for relocation assistance.</p>\n<p>Visit our PinFlex page to learn more about our working model.</p>\n<p>In-Office Requirement Statement:</p>\n<p>We recognize that the ideal environment for work is situational and may differ across departments.</p>\n<p>What this looks like day-to-day can vary based on the needs of each organization or role.</p>\n<p>This role will need to be in the office for in-person collaboration 1-2 times per quarter and therefore needs to be in a commutable distance from one of the following offices: PA or SF offices.</p>\n<p>#LI-HYBRID #LI-CL5</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_92f42798-2b7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/7747067","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$208,454-$364,795 USD","x-skills-required":["Java","Python","C/C++","Distributed Systems","Storage/Caching Systems","AI","Kubernetes","Memcached"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:04.357Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, US; Palo Alto, CA, US"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Python, C/C++, Distributed Systems, Storage/Caching Systems, AI, Kubernetes, Memcached","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":208454,"maxValue":364795,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_33821044-320"},"title":"Principal Engineer, Storage","description":"<p>We&#39;re looking for a Principal Engineer to play a key role in designing, building, and operating the data plane for our high-performance AI storage platform.</p>\n<p>You&#39;ll develop CoreWeave&#39;s storage systems by building reliable, scalable, and high-throughput solutions that power some of the largest and most innovative AI workloads in the world.</p>\n<p>This role involves close collaboration with teams across infrastructure, compute, and platform to ensure our storage services scale automatically and seamlessly while maximizing performance and reliability.</p>\n<p>About the role:</p>\n<ul>\n<li>Design and implement a highly scalable multi-tenant control plane that supports CoreWeave&#39;s growing AI storage and cloud infrastructure needs.</li>\n<li>Contribute to the development of exabyte-scale, S3-compatible object storage, distributed file system and integrate dedicated storage clusters into diverse customer environments.</li>\n<li>Work with technologies such as RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, and distributed filesystems to optimize storage performance and efficiency.</li>\n<li>Participate in efforts to improve the reliability, durability, and observability of our storage stack.</li>\n<li>Collaborate with operations teams to monitor, analyze, and optimize storage systems using telemetry, metrics, and dashboards to improve performance, latency, and resilience.</li>\n<li>Work cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack.</li>\n<li>Share your knowledge and mentor other engineers on best practices in building distributed, high-performance systems, especially focusing on low level storage details that improve performance and durability.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>Bachelor’s, Master’s, or PHD degree in Computer Science, Engineering, or a related field.</li>\n<li>8–10+ years of experience working in storage systems engineering.</li>\n<li>Strong hands-on experience with object storage, block storage or distributed filesystems in production environments.</li>\n<li>Proficiency in a systems programming language such as Go, C, or Rust.</li>\n<li>Solid understanding of cloud-native infrastructure, Kubernetes, and scalable system architecture.</li>\n<li>Strong debugging and problem-solving skills in distributed, high-performance environments.</li>\n<li>Clear communicator, able to work collaboratively across teams and share technical insights effectively.</li>\n<li>Familiarity with the trade offs between HDD and SSD based storage systems.</li>\n</ul>\n<p>The base salary range for this role is $206,000 to $303,000.</p>\n<p>In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.</p>\n<p>In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n<li>100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_33821044-320","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4646276006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$206,000 to $303,000","x-skills-required":["object storage","block storage","distributed filesystems","RDMA","GPU Direct Storage","RoCE","InfiniBand","SPDK","cloud-native infrastructure","Kubernetes","scalable system architecture","systems programming language","Go","C","Rust"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:53.363Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"object storage, block storage, distributed filesystems, RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, cloud-native infrastructure, Kubernetes, scalable system architecture, systems programming language, Go, C, Rust","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":206000,"maxValue":303000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0396ac1c-dad"},"title":"Senior Staff Engineer, Cloud Economics","description":"<p>Reddit is a community of communities. It&#39;s built on shared interests, passion, and trust, and is home to the most open and authentic conversations on the internet.</p>\n<p>The Ads Foundations organization is responsible for the technical backbone powering Ads Monetization at scale. Within this ecosystem, efficient resource utilization is critical.</p>\n<p>We are seeking a Senior Staff Engineer to serve as the Cloud Resources Technical Owner for the Ads Domain. You will be the primary engineering point of contact for the Senior Director in Ads and Cloud Operations/Resources (COR &amp; Opex) stakeholders.</p>\n<p><strong>Responsibilities</strong></p>\n<p>Technical Vision &amp; Strategy</p>\n<ul>\n<li>Define and drive the technical strategy for Cloud Resource management within Ad first, ensuring that cost accountability is built into the architecture of our systems.</li>\n<li>High-Fidelity Investment Modeling: Elevate cloud estimation from guesswork to a rigorous engineering discipline. You will lead the high-quality forecasting of new cloud investments and efficiency projects, designing data-driven models to validate technical ROI before builds happen</li>\n<li>Design and implement a roadmap for Cost Observability 2.0, moving beyond simple reporting to real-time, service/team-level spend attribution and automated anomaly detection.</li>\n</ul>\n<p>Engineering &amp; Tooling Leadership</p>\n<ul>\n<li>Design and build internal platforms that programmatically enforce PnL accountability. You will engineer (or collaborate with Core Infrastructure partners) to deliver the dashboards, alerts, and governance tools that every Ads team relies on to manage their cloud footprint.</li>\n<li>Architect automated frameworks for validating cost estimates and forecasting, replacing manual spreadsheets with data-driven software solutions.</li>\n</ul>\n<p>Scale &amp; Optimization</p>\n<ul>\n<li>Fight for observability by instrumenting deep telemetry into our cloud infrastructure. You will be hands-on in identifying inefficiencies (e.g., underutilized clusters, uncompressed data flows) and re-architecting critical paths for cost reduction.</li>\n<li>Lead the technical validation of vendor and 3rd-party tool integration, ensuring we extract maximum engineering value from every dollar spent.</li>\n</ul>\n<p>Cultural &amp; Technical Stewardship</p>\n<ul>\n<li>Act as a role model for the Ads domain and the wider company. You will set the standard for how engineering teams think about Cost as a Non Functional Requirement, eventually scaling these patterns to other domains.</li>\n<li>Partner with Finance and Engineering leadership to translate Cloud Spend into actionable engineering tasks (e.g., refactor Service X to use Spot instances).</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>10+ years of software engineering experience, with a strong focus on public cloud infrastructure (AWS/GCP/Azure) and large-scale distributed systems.</li>\n<li>Engineer-First Mindset: You are comfortable writing code (Go, Python, Java) to solve infrastructure problems. You don&#39;t just ask for a report; you build the API that generates it.</li>\n<li>Deep Cloud Expertise: You have mastery over Kubernetes, container orchestration, and cloud-native storage, understanding exactly how architectural choices impact the bottom line.</li>\n<li>Operational Excellence: Proven track record of building observability pipelines (Prometheus, Grafana, Datadog) that drive operational and financial alerts.</li>\n<li>Influential Leader: Skilled at driving clarity in ambiguous spaces. You can convince a Principal Engineer to refactor their service for cost efficiency because you can prove the technical and business value.</li>\n</ul>\n<p><strong>Bonus Points</strong></p>\n<ul>\n<li>Experience building custom FinOps tooling or internal developer platforms.</li>\n<li>Background in performance engineering or capacity planning for high-traffic ad tech environments.</li>\n<li>Contributions to open-source projects related to cloud efficiency or observability.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0396ac1c-dad","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit Inc.","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7628291","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$232,500-$325,500 USD","x-skills-required":["public cloud infrastructure","large-scale distributed systems","Kubernetes","container orchestration","cloud-native storage","observability pipelines","Prometheus","Grafana","Datadog"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:43.900Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"public cloud infrastructure, large-scale distributed systems, Kubernetes, container orchestration, cloud-native storage, observability pipelines, Prometheus, Grafana, Datadog","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":232500,"maxValue":325500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a02999d2-33b"},"title":"Staff Software Engineer - Backend","description":"<p>At Databricks, we are enabling data teams to solve the world&#39;s toughest problems by building and running the world&#39;s best data and AI infrastructure platform. As a software engineer with a backend focus, you will work with your team to build infrastructure and products for the Databricks platform at scale.</p>\n<p>The impact you&#39;ll have is significant, spanning many domains across our essential service platforms. You might work on challenges such as:</p>\n<ul>\n<li>Distributed systems, at-scale service architecture and monitoring, workflow orchestration, and developer experience.</li>\n</ul>\n<ul>\n<li>Delivering reliable and high-performance services and client libraries for storing and accessing humongous amounts of data on cloud storage backends, e.g., AWS S3, Azure Blob Store.</li>\n</ul>\n<ul>\n<li>Building reliable, scalable services, e.g., Scala, Kubernetes, and data pipelines, e.g., Spark, Databricks, to power the pricing infrastructure that serves millions of cluster-hours per day and develop product features that empower customers to easily view and control platform usage.</li>\n</ul>\n<p>What we look for in a candidate includes:</p>\n<ul>\n<li>A Bachelor&#39;s degree (or higher) in Computer Science, or a related field.</li>\n</ul>\n<ul>\n<li>7+ years of production-level experience in one of: Java, Scala, C++, or similar languages.</li>\n</ul>\n<ul>\n<li>Experience developing large-scale distributed systems.</li>\n</ul>\n<ul>\n<li>Experience working on a SaaS platform or with Service-Oriented Architectures.</li>\n</ul>\n<ul>\n<li>Good knowledge of SQL.</li>\n</ul>\n<p>Benefits at Databricks include comprehensive benefits and perks that meet the needs of all employees. For specific details on the benefits offered in your region, please click here.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a02999d2-33b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7984907002","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Scala","C++","SQL","distributed systems","at-scale service architecture and monitoring","workflow orchestration","developer experience","cloud storage backends","AWS S3","Azure Blob Store","Kubernetes","Spark","Databricks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:34.292Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Berlin, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Scala, C++, SQL, distributed systems, at-scale service architecture and monitoring, workflow orchestration, developer experience, cloud storage backends, AWS S3, Azure Blob Store, Kubernetes, Spark, Databricks"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_72ebb09d-b37"},"title":"Staff+ Software Engineer, Observability","description":"<p>We&#39;re seeking talented and experienced Software Engineers to join our Observability team within the Infrastructure organization. The Observability team owns the monitoring and telemetry infrastructure that every engineer and researcher at Anthropic depends on,from metrics and logging pipelines to distributed tracing, error analytics, alerting, and the dashboards and query interfaces that make it all actionable.</p>\n<p>As Anthropic scales its infrastructure across massive GPU, TPU, and Trainium clusters, the volume and complexity of operational data is growing by orders of magnitude. We&#39;re building next-generation observability systems,high-throughput ingest pipelines, cost-efficient columnar storage, unified query layers across signals, and agentic diagnostic tools,to ensure that engineers can detect, diagnose, and resolve issues in minutes rather than hours, even as the systems they operate become exponentially more complex.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build scalable telemetry ingest and storage pipelines for metrics, logs, traces, and error data across Anthropic&#39;s multi-cluster infrastructure</li>\n<li>Own and evolve core observability platforms, driving migrations and architectural improvements that improve reliability, reduce cost, and scale with organisational growth</li>\n<li>Build instrumentation libraries, SDKs, and integrations that make it easy for engineering teams to emit high-quality telemetry from their services</li>\n<li>Drive alerting and SLO infrastructure that enables teams to define, monitor, and respond to reliability targets with minimal noise</li>\n<li>Reduce mean time to detection and resolution by building cross-signal correlation, unified query interfaces, and AI-assisted diagnostic tooling</li>\n<li>Partner with Research, Inference, Product, and Infrastructure teams to ensure observability solutions meet the unique needs of each organisation</li>\n</ul>\n<p>You May Be a Good Fit If You:</p>\n<ul>\n<li>Have 10+ years of relevant industry experience building and operating large-scale observability or monitoring infrastructure</li>\n<li>Have deep experience with at least one observability signal area (metrics, logging, tracing, or error analytics) and familiarity with the others</li>\n<li>Understand high-throughput data pipelines, columnar storage engines, and the tradeoffs involved in ingesting and querying telemetry data at scale</li>\n<li>Have experience operating or building on top of observability platforms such as Prometheus, Grafana, ClickHouse, OpenTelemetry, or similar systems</li>\n<li>Have strong proficiency in at least one of Python, Rust, or Go</li>\n<li>Have excellent communication skills and enjoy partnering with internal teams to improve their operational visibility and incident response capabilities</li>\n<li>Are excited about building foundational infrastructure and are comfortable working independently on ambiguous, high-impact technical challenges</li>\n</ul>\n<p>Strong Candidates May Also Have:</p>\n<ul>\n<li>Experience operating metrics systems at very high cardinality (hundreds of millions of active time series or more)</li>\n<li>Experience with log storage migrations or operating columnar databases (ClickHouse, BigQuery, or similar) for analytics workloads</li>\n<li>Experience with OpenTelemetry instrumentation, collector pipelines, and tail-based sampling strategies</li>\n<li>Experience building or operating alerting platforms, on-call tooling, or SLO frameworks at scale</li>\n<li>Experience with Kubernetes-native monitoring, eBPF-based observability, or continuous profiling</li>\n<li>Interest in applying AI/LLMs to operational workflows such as automated root cause analysis, anomaly detection, or intelligent alerting</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_72ebb09d-b37","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5139910008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["observability","monitoring","telemetry","metrics","logging","tracing","error analytics","alerting","SLO infrastructure","cross-signal correlation","unified query interfaces","AI-assisted diagnostic tooling","Python","Rust","Go","Prometheus","Grafana","ClickHouse","OpenTelemetry"],"x-skills-preferred":["high-throughput data pipelines","columnar storage engines","operating system administration","cloud computing","containerization","DevOps"],"datePosted":"2026-04-18T15:51:29.494Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"observability, monitoring, telemetry, metrics, logging, tracing, error analytics, alerting, SLO infrastructure, cross-signal correlation, unified query interfaces, AI-assisted diagnostic tooling, Python, Rust, Go, Prometheus, Grafana, ClickHouse, OpenTelemetry, high-throughput data pipelines, columnar storage engines, operating system administration, cloud computing, containerization, DevOps","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_22ff82ac-40b"},"title":"Software Engineer, Research Data Platform","description":"<p>We&#39;re looking for engineers who love working directly with users and who excel at building data products. The Research Data Platform team builds the tools that Anthropic&#39;s researchers use every day to manage, query, and analyze the data that goes into training and evaluating frontier models.</p>\n<p>As a software engineer on this team, you will:</p>\n<ul>\n<li>Build and operate data pipelines that extract data from research training runs and land it in storage systems that are easy and fast to query</li>\n<li>Work closely with researchers to design and build APIs, libraries, and web interfaces that support data management, exploration, and analysis</li>\n<li>Develop dataset management, data cataloging, and provenance tooling that researchers use in their day-to-day work</li>\n<li>Embed with research teams to understand their workflows, identify high-leverage tooling opportunities, and ship solutions quickly</li>\n<li>Collaborate with adjacent teams to build on existing systems rather than reinventing them</li>\n</ul>\n<p>You may be a good fit if you have significant software engineering experience, particularly building data-intensive applications or internal tooling. You should enjoy working directly with users, gathering requirements iteratively, and shipping things that get adopted. You should also be results-oriented, with a bias towards flexibility and impact.</p>\n<p>Strong candidates may also have experience with large-scale ETL, columnar storage formats, and query engines, high-volume time series data, data cataloging, lineage, or metadata management systems, ML experiment tracking or metrics platforms, complex data visualization, and full-stack web application development.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_22ff82ac-40b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5191226008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["software engineering","data-intensive applications","internal tooling","data pipelines","storage systems","APIs","libraries","web interfaces","dataset management","data cataloging","provenance tooling","research workflows","adjacent teams"],"x-skills-preferred":["large-scale ETL","columnar storage formats","query engines","high-volume time series data","lineage","metadata management systems","ML experiment tracking","metrics platforms","complex data visualization","full-stack web application development"],"datePosted":"2026-04-18T15:51:29.293Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"software engineering, data-intensive applications, internal tooling, data pipelines, storage systems, APIs, libraries, web interfaces, dataset management, data cataloging, provenance tooling, research workflows, adjacent teams, large-scale ETL, columnar storage formats, query engines, high-volume time series data, lineage, metadata management systems, ML experiment tracking, metrics platforms, complex data visualization, full-stack web application development","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0ae48270-bef"},"title":"Senior Software Engineer, Storage Engineer","description":"<p>The Storage Engine Organisation at CoreWeave is responsible for the product capabilities and data plane function of CoreWeave&#39;s managed storage products.</p>\n<p>We build reliable, scalable storage solutions with segment leading performance. Storage engine works with engineering teams across infrastructure, compute, and platform to ensure our storage services meet the needs of the world&#39;s most demanding AI workloads.</p>\n<p>The role involves designing and implementing distributed storage solutions to support scaling data-intensive AI workloads, contributing to the development of exabyte-scale, S3-compatible object storage, and integrating dedicated storage clusters into diverse customer environments.</p>\n<p>Key responsibilities include working with technologies such as RDMA, GPU Direct Storage, and distributed filesystems protocols like NFS or FUSE to optimise storage performance and efficiency, participating in efforts to improve the reliability, durability, and observability of our storage stack, collaborating with operations teams to monitor, troubleshoot, and improve storage systems in production environments, and helping develop metrics and dashboards to provide visibility into storage performance and health.</p>\n<p>The ideal candidate will have a strong background in storage systems engineering or infrastructure, with experience working with object storage or distributed filesystems in production environments, proficiency in a systems programming language like Go, C, or Rust, and familiarity with storage observability tools and telemetry pipelines.</p>\n<p>As a senior software engineer, you will be responsible for designing, developing, and deploying scalable and efficient storage solutions, working closely with cross-functional teams to ensure seamless integration with other components of the platform, and mentoring junior engineers to help them grow in their roles.</p>\n<p>If you are passionate about building high-performance storage solutions and have a strong background in software engineering, we encourage you to apply for this exciting opportunity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0ae48270-bef","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4643524006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$139,000 to $204,000","x-skills-required":["Storage systems engineering","Infrastructure","Object storage","Distributed filesystems","RDMA","GPU Direct Storage","NFS","FUSE","Systems programming languages (Go, C, Rust)","Storage observability tools","Telemetry pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:26.395Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ/ New York , NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Storage systems engineering, Infrastructure, Object storage, Distributed filesystems, RDMA, GPU Direct Storage, NFS, FUSE, Systems programming languages (Go, C, Rust), Storage observability tools, Telemetry pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":139000,"maxValue":204000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ff4d3a91-b20"},"title":"Principal Engineer - Perf and Benchmarking","description":"<p>We&#39;re looking for a Principal Engineer to be the technical lead of CoreWeave&#39;s Benchmarking &amp; Performance team. You will be responsible for our planet-scale performance data warehouse: Ingesting, storing, transforming and analyzing performance events in all the data centers across our global infrastructure.</p>\n<p>You will also be an integral part of achieving industry-leading end-to-end performance benchmarking publications: If MLPerf (Training &amp; Inference), Working closely with NVIDIA (Megatron-LM, TensorRT-LLM &amp; DGX cloud) and the open-source community (llm-d, vLLM and all popular ML frameworks) speak to you, come help us demonstrate CoreWeave&#39;s performance reliability leadership in the field.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Strategy &amp; Leadership - Define the multi-year benchmarking strategy and roadmap; prioritize models/workloads (LLMs, diffusion, vision, speech) and hardware tiers. Build, lead, and mentor a high-performing team of performance engineers and data analysts. Establish governance for claims: documented methodologies, versioning, reproducibility, and audit trails.</li>\n</ul>\n<ul>\n<li>Perf Ownership - Lead end-to-end MLPerf Inference and Training submissions: workload selection, cluster planning, runbooks, audits, and result publication. Coordinate optimization tracks with NVIDIA (CUDA, cuDNN, TensorRT/TensorRT-LLM, Triton, NCCL) to hit competitive results; drive upstream fixes where needed.</li>\n</ul>\n<ul>\n<li>Internal Latency &amp; Throughput Benchmarks - Design a Kubernetes-native, repeatable benchmarking service that exercises CoreWeave stacks across SUNK (Slurm on Kubernetes), Kueue, and Kubeflow pipelines. Measure and report p50/p95/p99 latency, jitter, tokens/s, time-to-first-token, cold-start/warm-start, and cost-per-token/request across models, precisions (BF16/FP8/FP4), batch sizes, and GPU types. Maintain a corpus of representative scenarios (streaming, batch, multi-tenant) and data sets; automate comparisons across software releases and hardware generations.</li>\n</ul>\n<ul>\n<li>Tooling &amp; Automation - Build CI/CD pipelines and K8s controllers/operators to schedule benchmarks at scale; integrate with observability stacks (Prometheus, Grafana, OpenTelemetry) and results warehouses. Implement supply-chain integrity for benchmark artifacts (SBOMs, Cosign signatures).</li>\n</ul>\n<ul>\n<li>Cross-functional &amp; Community - Partner with NVIDIA, key ISVs, and OSS projects (vLLM, Triton, KServe, PyTorch/DeepSpeed, ONNX Runtime) to co-develop optimizations and upstream improvements. Support Sales/SEs with authoritative numbers for RFPs and competitive evaluations; brief analysts and press with rigorous, defensible data.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>10+ years building distributed systems or HPC/cloud services, with deep expertise on large-scale ML training or similar high-performance workloads.</li>\n</ul>\n<ul>\n<li>Proven track record of architecting or building planet-scale data systems (e.g., telemetry platforms, observability stacks, cloud data warehouses, large-scale OLAP engines).</li>\n</ul>\n<ul>\n<li>Deep understanding of GPU performance (CUDA, NCCL, RDMA, NVLink/PCIe, memory bandwidth), model-server stacks (Triton, vLLM, TensorRT-LLM, TorchServe), and distributed training frameworks (PyTorch FSDP/DeepSpeed/Megatron-LM).</li>\n</ul>\n<ul>\n<li>Proficient with Kubernetes and ML control planes; familiarity with SUNK, Kueue, and Kubeflow in production environments.</li>\n</ul>\n<ul>\n<li>Excellent communicator able to interface with executives, customers, auditors, and OSS communities.</li>\n</ul>\n<p><strong>Nice to have</strong></p>\n<ul>\n<li>Experience with time-series databases, log-structured merge trees (LSM), or custom storage engine development.</li>\n</ul>\n<ul>\n<li>Experience running MLPerf submissions (Inference and/or Training) or equivalent audited benchmarks at scale.</li>\n</ul>\n<ul>\n<li>Contributions to MLPerf, Triton, vLLM, PyTorch, KServe, or similar OSS projects.</li>\n</ul>\n<ul>\n<li>Experience benchmarking multi-region fleets and large clusters (thousands of GPUs).</li>\n</ul>\n<ul>\n<li>Publications/talks on ML performance, latency engineering, or large-scale benchmarking methodology.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ff4d3a91-b20","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4627302006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$206,000 to $333,000","x-skills-required":["Distributed systems","HPC/cloud services","Large-scale ML training","GPU performance","Model-server stacks","Distributed training frameworks","Kubernetes","ML control planes","Time-series databases","Log-structured merge trees","Custom storage engine development"],"x-skills-preferred":["MLPerf submissions","Audited benchmarks","Contributions to OSS projects","Benchmarking multi-region fleets","Large clusters","Publications/talks on ML performance"],"datePosted":"2026-04-18T15:51:17.448Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Distributed systems, HPC/cloud services, Large-scale ML training, GPU performance, Model-server stacks, Distributed training frameworks, Kubernetes, ML control planes, Time-series databases, Log-structured merge trees, Custom storage engine development, MLPerf submissions, Audited benchmarks, Contributions to OSS projects, Benchmarking multi-region fleets, Large clusters, Publications/talks on ML performance","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":206000,"maxValue":333000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b00b781c-eba"},"title":"Senior Software Engineer - Database Engine Internals","description":"<p>We&#39;re seeking a Senior Software Engineer to join our team in designing next-generation systems for database engine internals. As part of this multi-year journey, you&#39;ll drive requirements clarity and design decisions for ambiguous problems. Your responsibilities will include producing technical design documents and project plans, developing new features, mentoring junior engineers, testing and rolling out to production, and monitoring.</p>\n<p>Our ideal candidate has a passion for database systems, storage systems, distributed systems, language design, or performance optimisation. They should be comfortable working towards a multi-year vision with incremental deliverables and be customer-oriented with a focus on having an impact. A minimum of 5 years of experience working in a related system is required, with a PhD in databases or distributed systems being optional.</p>\n<p>In return, we offer a comprehensive benefits package and a commitment to diversity and inclusion. If you&#39;re excited about the opportunity to join our team and contribute to the development of next-generation database systems, please apply.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b00b781c-eba","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8012809002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimisation","Apache Spark","Delta Lake","MLflow"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:15.118Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Belgrade, Serbia"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimisation, Apache Spark, Delta Lake, MLflow"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f2c6f765-eca"},"title":"Staff Engineer, Storage Control Plane","description":"<p>We&#39;re looking for a Staff Storage Engineer to play a key role in designing, building, and operating the control plane for our high-performance AI storage platform. You&#39;ll help evolve CoreWeave&#39;s storage systems by building reliable, scalable, and high-throughput solutions that power some of the largest and innovative AI workloads in the world.</p>\n<p>This role involves close collaboration with teams across infrastructure, compute, and platform to ensure our storage services scale automatically and seamlessly while maximizing performance and reliability.</p>\n<p>About the role:</p>\n<ul>\n<li>Design and implement a highly scalable multi-tenant control plane that supports CoreWeave&#39;s growing AI storage and cloud infrastructure needs.</li>\n</ul>\n<ul>\n<li>Contribute to the development of exabyte-scale, S3-compatible object storage, distributed file system and integrate dedicated storage clusters into diverse customer environments.</li>\n</ul>\n<ul>\n<li>Work with technologies such as RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, and distributed filesystems to optimize storage performance and efficiency.</li>\n</ul>\n<ul>\n<li>Participate in efforts to improve the reliability, durability, and observability of our storage stack.</li>\n</ul>\n<ul>\n<li>Collaborate with operations teams to monitor, analyze, and optimize storage systems using telemetry, metrics, and dashboards to improve performance, latency, and resilience.</li>\n</ul>\n<ul>\n<li>Work cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack.</li>\n</ul>\n<ul>\n<li>Share your knowledge and mentor other engineers on best practices in building distributed, high-performance systems.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>Bachelor&#39;s or Master&#39;s degree in Computer Science, Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>10+ years of experience working in storage systems engineering or infrastructure.</li>\n</ul>\n<ul>\n<li>Strong hands-on experience with object storage or distributed filesystems in production environments.</li>\n</ul>\n<ul>\n<li>Experience with one or more storage protocols (e.g. S3, NFS) and file systems such as Ceph, DAOS, or similar.</li>\n</ul>\n<ul>\n<li>Proficiency in a systems programming language such as Go, C++, or Rust.</li>\n</ul>\n<ul>\n<li>Familiarity with storage observability tools and telemetry pipelines (e.g., ClickHouse, Prometheus, Grafana).</li>\n</ul>\n<ul>\n<li>Solid understanding of cloud-native infrastructure, Kubernetes, and scalable system architecture.</li>\n</ul>\n<ul>\n<li>Strong debugging and problem-solving skills in distributed, high-performance environments.</li>\n</ul>\n<ul>\n<li>Clear communicator, able to work collaboratively across teams and share technical insights effectively.</li>\n</ul>\n<p>Wondering if you&#39;re a good fit? We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren&#39;t a 100% skill or experience match. Here are a few qualities we&#39;ve found compatible with our team. If some of this describes you, we&#39;d love to talk.</p>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n</ul>\n<ul>\n<li>Act Like an Owner</li>\n</ul>\n<ul>\n<li>Empower Employees</li>\n</ul>\n<ul>\n<li>Deliver Best-in-Class Client Experiences</li>\n</ul>\n<ul>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!</p>\n<p>The base salary range for this role is $165,000 to $242,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we&#39;ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location. In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n</ul>\n<ul>\n<li>100% paid for by CoreWeave</li>\n</ul>\n<ul>\n<li>Company-paid Life Insurance</li>\n</ul>\n<ul>\n<li>Voluntary supplemental life insurance</li>\n</ul>\n<ul>\n<li>Short and long-term disability insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Account</li>\n</ul>\n<ul>\n<li>Health Savings Account</li>\n</ul>\n<ul>\n<li>Tuition Reimbursement</li>\n</ul>\n<ul>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n</ul>\n<ul>\n<li>Mental Wellness Benefits through Spring Health</li>\n</ul>\n<ul>\n<li>Family-Forming support provided by Carrot</li>\n</ul>\n<ul>\n<li>Paid Parental Leave</li>\n</ul>\n<ul>\n<li>Flexible, full-service childcare support with Kinside</li>\n</ul>\n<ul>\n<li>401(k) with a generous employer match</li>\n</ul>\n<ul>\n<li>Flexible PTO</li>\n</ul>\n<ul>\n<li>Catered lunch each day in our office and data center locations</li>\n</ul>\n<ul>\n<li>A casual work environment</li>\n</ul>\n<ul>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p>California Consumer Privacy Act - California applicants only</p>\n<p>CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information. As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com.</p>\n<p>Export Control Compliance</p>\n<p>This position requires access to export controlled information. To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without a required export authorization, or (C) eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency. CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f2c6f765-eca","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4669836006","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["object storage","distributed filesystems","storage protocols","file systems","cloud-native infrastructure","Kubernetes","scalable system architecture","systems programming language","Go","C++","Rust","storage observability tools","telemetry pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:06.353Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA / Dallas, TX"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"object storage, distributed filesystems, storage protocols, file systems, cloud-native infrastructure, Kubernetes, scalable system architecture, systems programming language, Go, C++, Rust, storage observability tools, telemetry pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1b5b24ef-246"},"title":"Engineering Manager II, Programmatic Offsite Ads","description":"<p>About Pinterest</p>\n<p>We&#39;re on a mission to bring everyone the inspiration to create a life they love, and that starts with the people behind the product. Discover a career where you ignite innovation for millions, transform passion into growth opportunities, celebrate each other&#39;s unique experiences and embrace the flexibility to do your best work.</p>\n<p>Creating a career you love? It&#39;s Possible. At Pinterest, AI isn&#39;t just a feature, it&#39;s a powerful partner that augments our creativity and amplifies our impact, and we’re looking for candidates who are excited to be a part of that.</p>\n<p>To get a complete picture of your experience and abilities, we’ll explore your foundational skills and how you collaborate with AI. Through our interview process, what matters most is that you can always explain your approach, showing us not just what you know, but how you think.</p>\n<p>You can read more about our AI interview philosophy and how we use AI in our recruiting process here.</p>\n<p>Job Summary</p>\n<p>We’re seeking a talented Manager II, Engineering to take on a leadership role within the Programmatic Offsite Ads team. You will lead critical efforts to define, build, and evolve the ad features which power Pinterest’s ads business through off-platform supply partnerships.</p>\n<p>Responsibilities</p>\n<p>In this pivotal role, you will take on the challenge of defining and executing the offsite ads strategy for programmatic ads at Pinterest.</p>\n<p>Own the end-to-end strategy and roadmap for driving programmatic off-platform ads delivery, driving high-quality outcomes which meet advertiser expectations.</p>\n<p>Partner closely with Product, Design, Research, Sales, Policy, and the broader Monetization org to define new product features, and advertising experiences that balance user delight, advertiser outcomes, and platform integrity.</p>\n<p>Lead experimentation and optimization of advertising campaigns, using A/B testing and rigorous measurement (e.g., viewability, engagement, conversion, advertiser performance, user sentiment) to drive continuous improvement.</p>\n<p>Work with external supply partners to ensure our off-platform ads are well-supported in the programmatic ecosystem, and that Pinterest’s creatives adhere to performance standards.</p>\n<p>Collaborate with serving, infra, and ML teams to ensure that programmatic ads are backed by robust infrastructure, measurement, and policies.</p>\n<p>Lead mission-critical initiatives involving 8-10 engineers across backend and frontend stacks, and directly influence their day-to-day work through mentorship, coaching, and clear technical direction.</p>\n<p>Build and maintain a culture of inclusivity, craft, and operational excellence within the Programmatic Offsite Ads team.</p>\n<p>Collaborate with stakeholders and partner teams across the organization to architect data lake storage and metadata management technologies to unlock big data and ML/AI innovations.</p>\n<p>Use AI to accelerate analysis, iteration, experimentation and time to market while applying judgment and verification to ensure correctness and quality.</p>\n<p>Requirements</p>\n<p>BS (or higher) degree in Computer Science, or a related field.</p>\n<p>2-3+ years of relevant engineering management experience.</p>\n<p>3-4+ years of relevant industry experience within the ads domain.</p>\n<p>Experience designing or delivering high scale, real time distributed systems.</p>\n<p>Working knowledge of programmatic advertising and OpenRTB (DSPs/SSPs, auctions, targeting, measurement), and experience partnering with external platforms.</p>\n<p>Proven track record partnering with Product and Design to define new product features, run experiments, and use data to iterate on performance outcomes.</p>\n<p>Rich experience working cross-functionally to drive alignment, oversee execution, and secure deliverables across Product, Design, ML, Infra, Sales, and external partners.</p>\n<p>Build storage capabilities that efficiently support large-scale ML/AI workloads, including high-throughput data access, schema evolution, and large-scale column backfills.</p>\n<p>Demonstrated ability to use AI to improve speed and quality in your day-to-day workflow for relevant outputs.</p>\n<p>High integrity and ownership: you protect sensitive data, avoid over-reliance on AI, and remain accountable for final decisions and deliverables.</p>\n<p>Experience mentoring, guiding, and upleveling engineers, including senior ICs.</p>\n<p>Strong communication skills and the ability to articulate product strategy and tradeoffs to both technical and non-technical stakeholders.</p>\n<p>Strong commitment to building inclusive teams and fostering a sense of belonging.</p>\n<p>In-Office Requirement Statement:</p>\n<p>We let the type of work you do guide the collaboration style. That means we&#39;re not always working in an office, but we continue to gather for key moments of collaboration and connection.</p>\n<p>This role will need to be in the office for in-person collaboration [1 time per week] and therefore needs to be in a commutable distance from one of the following offices: San Francisco.</p>\n<p>Relocation Statement:</p>\n<p>This position is not eligible for relocation assistance. Visit our PinFlex page to learn more about our working model.</p>\n<p>#LI-HYBRID #LI-KBF</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1b5b24ef-246","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/7494773","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$177,185-$364,795 USD","x-skills-required":["Computer Science","Engineering Management","Programmatic Advertising","OpenRTB","Distributed Systems","Data Lake Storage","Metadata Management","AI","Machine Learning"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:52.945Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, US; Palo Alto, CA, US"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Computer Science, Engineering Management, Programmatic Advertising, OpenRTB, Distributed Systems, Data Lake Storage, Metadata Management, AI, Machine Learning","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":177185,"maxValue":364795,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a027f462-69a"},"title":"Senior Software Developer - Storage Engine - Elasticsearch","description":"<p>We&#39;re looking for a Senior Software Developer to join the team that contributes to improving our storage efficiency for metrics, logs, and other types of data. As a software engineer in the team, you will work on different initiatives, such as enhancing current logging solutions to ensure that logging data is always accepted and persisted, advancing our current metrics processing capabilities to ensure massive and seamless adoption by our customers, and improving storage efficiency across the board. You&#39;ll also be extending the logic for efficiently querying and aggregating the stored data, taking their storage layout and ordering into account.</p>\n<p>Our company is distributed by intention. We hire the best engineers we can find wherever they are, whoever they are. We collaborate across continents every day over email, GitHub, Zoom, and Slack. At our best, we write fast, scalable, intuitive, and high-quality software. We believe that the best way to do that is to empower individual engineers, code review every change, decide big things by consensus, and strive for incremental improvements.</p>\n<p>As a Senior Software Developer, you will:</p>\n<ul>\n<li>Work with a globally distributed team of experienced engineers focused on data storage mechanisms and query capabilities of Elasticsearch.</li>\n<li>Be an expert in the storage engine area, and everyone will turn to you when they have a question about it. You&#39;ll improve those areas based on your questions and your instincts.</li>\n<li>Be a full-time Elasticsearch contributor, building data-intensive new features, fixing intriguing bugs, and increasing the testing coverage, all while making the code easier to understand.</li>\n<li>Design and implement advanced algorithms and data structures, often working at the system and hardware level. You’ll also engage with our global community for triaging and resolving issues and pull requests.</li>\n</ul>\n<p>We&#39;re looking for someone with strong core Java skills and an excellent understanding of concurrent and parallel programming principles. You should have an excellent background in applied data processing (data structures, algorithms) and be familiar with storage systems and low-level abstractions in OS. You should also be able to work with a high level of autonomy and be able to take on projects and guide them from beginning to end.</p>\n<p>This role does not have a variable compensation component. The typical starting salary range for new hires in this role is $128,300-$203,000 CAD. This role is currently eligible to participate in Elastic&#39;s stock program. Our total rewards package also includes a company-matched Registered Retirement Savings Plan (RRSP) with dollar-for-dollar matching up to 6% of eligible earnings, along with a range of other benefits offered with a holistic emphasis on employee well-being.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a027f462-69a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Elastic","sameAs":"https://www.elastic.co/","logo":"https://logos.yubhub.co/elastic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/elastic/jobs/7592630","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$128,300-$203,000 CAD","x-skills-required":["Java","Concurrent and parallel programming principles","Data structures and algorithms","Storage systems and low-level abstractions in OS"],"x-skills-preferred":["Elasticsearch","Solr","PostgreSQL","MongoDB","Cassandra"],"datePosted":"2026-04-18T15:50:33.450Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Concurrent and parallel programming principles, Data structures and algorithms, Storage systems and low-level abstractions in OS, Elasticsearch, Solr, PostgreSQL, MongoDB, Cassandra","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":128300,"maxValue":203000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5b6f9322-a9a"},"title":"Staff Engineer, Storage Engine","description":"<p>CoreWeave is seeking a Staff Engineer, Storage Engine to join their team. The successful candidate will design and implement distributed storage solutions to support scaling data-intensive AI workloads. They will contribute to the development of exabyte-scale, S3-compatible object storage and integrate dedicated storage clusters into diverse customer environments.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Designing and implementing distributed storage solutions to support scaling data-intensive AI workloads</li>\n<li>Contributing to the development of exabyte-scale, S3-compatible object storage</li>\n<li>Integrating dedicated storage clusters into diverse customer environments</li>\n<li>Working with technologies such as RDMA, GPU Direct Storage, and distributed filesystems protocols such as NFS or FUSE to optimize storage performance and efficiency</li>\n<li>Leading efforts to improve the reliability, durability, security, and observability of the storage stack</li>\n<li>Collaborating with operations teams to monitor, troubleshoot, and improve storage systems in production environments</li>\n<li>Setting the bar for developing metrics and dashboards to provide visibility into storage performance and health</li>\n<li>Analyzing telemetry and system data to drive improvements in throughput, latency, and resilience</li>\n<li>Working cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack</li>\n<li>Sharing knowledge and mentoring other engineers on best practices in building distributed, high-performance systems</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>Bachelor&#39;s, Master&#39;s, or PhD degree in Computer Science, Engineering, or a related field</li>\n<li>8-10+ years of experience working in storage systems engineering or infrastructure</li>\n<li>Strong hands-on experience with object storage or distributed filesystems in production environments</li>\n<li>Experience with one or more storage protocols (e.g. S3, NFS) and file systems such as Ceph, DAOS, or similar</li>\n<li>Proficiency in a systems programming language such as Go, C, or Rust</li>\n<li>Proficiency leveraging AI tools to augment software development</li>\n<li>Familiarity with storage observability tools and telemetry pipelines (e.g., ClickHouse, Prometheus, Grafana)</li>\n<li>Experience working with cloud-native infrastructure, Kubernetes, and scalable system architectures</li>\n</ul>\n<p>The base salary range for this role is $188,000 to $275,000.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5b6f9322-a9a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4612047006","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$188,000 to $275,000","x-skills-required":["distributed storage","object storage","S3-compatible object storage","RDMA","GPU Direct Storage","distributed filesystems protocols","NFS","FUSE","storage performance and efficiency","reliability","durability","security","observability","telemetry","system data","throughput","latency","resilience","cloud-native infrastructure","Kubernetes","scalable system architectures"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:33.024Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"distributed storage, object storage, S3-compatible object storage, RDMA, GPU Direct Storage, distributed filesystems protocols, NFS, FUSE, storage performance and efficiency, reliability, durability, security, observability, telemetry, system data, throughput, latency, resilience, cloud-native infrastructure, Kubernetes, scalable system architectures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":188000,"maxValue":275000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a7d0cf0f-a3a"},"title":"Senior Engineer- Data Platforms","description":"<p>The Data Platform Team serves as the experts on managing data infrastructure for CoreWeave. Our data infrastructure includes managed databases, data ingestion, data flow, data lakes, and other data retrieval for CoreWeave and its customers.</p>\n<p>We are seeking senior software engineers with specialization in database and stream processing who can help us fulfill the goal of our global datastore strategy and establish communication models for our data flow. This individual will work with a team of mixed skilled engineers and have the opportunity to work on the full range of rewarding challenges that come with the business of building a cloud in a communicative, supportive, and high-performing environment.</p>\n<p>As a member of the Data Platform Team you will have the opportunity to:</p>\n<ul>\n<li>Design and implement the platform to deliver data to teams with a focus on providing managed solutions through APIs</li>\n<li>Participate in operations and scaling of relational data platforms</li>\n<li>Develop a stream processing architecture and solve for scalability and reliability</li>\n<li>Improve the performance, security, reliability, and scalability of our data platforms and related services, and participate in the team’s on-call rotation</li>\n<li>Establish guidelines and guard rails for data access and storage for stakeholder teams</li>\n<li>Ensure compliance with standards for data protection regulation</li>\n<li>Grow, change, invest in your teammates, be invested-in, share your ideas, listen to others, be curious, have fun, and, above all, be yourself</li>\n</ul>\n<p>The ideal candidate will have 5+ years of experience in a software or infrastructure engineering industry, with experience operating services in production and at scale and familiarity with reliability engineering concepts such as different types of testing, progressive deployments, error budgets, observability, and fault-tolerant design.</p>\n<p>The base salary range for this role is $175,000 to $210,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a7d0cf0f-a3a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4562276006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $210,000","x-skills-required":["database and stream processing","managed databases","data ingestion","data flow","data lakes","APIs","operational experience","reliability engineering","testing","progressive deployments","error budgets","observability","fault-tolerant design"],"x-skills-preferred":["Kubernetes","Go","Linux distributions","shell scripting","Linux storage and networking stacks"],"datePosted":"2026-04-18T15:50:18.835Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bellevue, WA / Sunnyvale, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database and stream processing, managed databases, data ingestion, data flow, data lakes, APIs, operational experience, reliability engineering, testing, progressive deployments, error budgets, observability, fault-tolerant design, Kubernetes, Go, Linux distributions, shell scripting, Linux storage and networking stacks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":210000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_30e11492-d62"},"title":"Software Engineer, Safeguards Infrastructure","description":"<p>We are looking for software engineers to help build the foundational pieces for safety, oversight and intervention mechanisms of our AI systems. As a software engineer on the Safeguards team, you will work to monitor models, prevent misuse, and ensure user well-being. This role will focus on building systems to detect unwanted model behaviors and prevent disallowed use of models. You will apply your technical skills to uphold our principles of safety, transparency, and oversight.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Develop the foundational systems which power Safeguards, including infrastructure for data storage and management, metric and evaluation systems, and tooling for human and agentic review.</li>\n<li>Ensure the day-to-day running of Safeguards systems and hold a high operational bar which serves both safety and customers while reducing the amount of human intervention and oversight required.</li>\n<li>Build robust and reliable multi-layered defenses for real-time improvement of safety mechanisms that work at scale</li>\n</ul>\n<p>You may be a good fit if you have:</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Software Engineering or comparable experience</li>\n<li>4-10+ years of experience in a software engineering position</li>\n<li>Proficiency in Python</li>\n<li>Ability to work across the stack</li>\n<li>Strong communication skills and ability to explain complex technical concepts to non-technical stakeholders</li>\n</ul>\n<p>Strong candidates may also:</p>\n<ul>\n<li>Have experience building trust and safety, anti-spam, fraud or abuse detection and mitigation mechanisms and interventions for AI/ML systems</li>\n<li>Have experience building metrics and measurement systems or data and privacy management systems</li>\n<li>Have worked closely with operational teams to build custom internal tooling</li>\n<li>Be proficient in TypeScript or Rust</li>\n<li>Have experience with Claude Code or similar agentic coding tools</li>\n</ul>\n<p>The annual compensation range for this role is £255,000-£325,000 GBP.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_30e11492-d62","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5074908008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"£255,000-£325,000 GBP","x-skills-required":["Python","Software Engineering","Data Storage and Management","Metric and Evaluation Systems","Tooling for Human and Agentic Review"],"x-skills-preferred":["TypeScript","Rust","Claude Code","Agentic Coding Tools"],"datePosted":"2026-04-18T15:50:15.126Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Software Engineering, Data Storage and Management, Metric and Evaluation Systems, Tooling for Human and Agentic Review, TypeScript, Rust, Claude Code, Agentic Coding Tools","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":255000,"maxValue":325000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f509d407-0b5"},"title":"Senior Security Engineer, Incident Response","description":"<p>You will be an individual contributor on the security Incident Response (IR) team at Databricks, reporting to the regional IR manager. Your responsibilities will include conducting security analysis and forensics, responding to high-priority alerts, and contributing to automations and agentic capabilities. You will be a security multiplier and help the team scale security incident response at Databricks.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Respond to incidents as part of a distributed 24x7 operations and on-call schedule.</li>\n<li>Triage and respond to security events and alerts, ensuring quick and effective containment.</li>\n<li>Contribute to security investigations, conducting analysis and forensics across a range of data sources to determine the timeline and impact of security events.</li>\n<li>Build automations, including leveraging AI and agentic platforms, to deliver autonomous capabilities, expedite your work, and scale the impact of the team.</li>\n<li>Communicate technical decisions through design docs and tech talks, and mentor junior security responders via security guidance, design reviews, and code reviews.</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>Bachelor&#39;s Degree AND 4+ years experience in Incident Response work OR Master&#39;s Degree AND 2+ years experience.</li>\n<li>Strong cloud security background in at least 1 of AWS, GCP, or Azure, and working knowledge of the others.</li>\n<li>Knowledge of AI/LLM and agentic capabilities, including effective prompting and use of MCP, agents, and agent skills.</li>\n<li>Broad security subject matter expertise.</li>\n<li>Expertise in few core IR skills (DFIR, Reverse Engineering, Traditional Network Security, Storage and access security, Sandboxing, Compute security, etc.).</li>\n<li>Experience with Enterprise Security and SaaS applications.</li>\n<li>Working knowledge of a SIEM and SOAR.</li>\n<li>Experience building Incident Response Tooling and scripting language skills.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f509d407-0b5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8026632002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["cloud security","AWS","GCP","Azure","AI/LLM","agentic capabilities","security subject matter expertise","DFIR","Reverse Engineering","Traditional Network Security","Storage and access security","Sandboxing","Compute security","Enterprise Security","SaaS applications","SIEM","SOAR","Incident Response Tooling","scripting language skills"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:52.742Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Amsterdam, Netherlands; Berlin, Germany; London, United Kingdom; Remote - Denmark; Remote - France; Remote - Germany; Remote - Italy; Remote - Spain; Remote - Sweden"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"cloud security, AWS, GCP, Azure, AI/LLM, agentic capabilities, security subject matter expertise, DFIR, Reverse Engineering, Traditional Network Security, Storage and access security, Sandboxing, Compute security, Enterprise Security, SaaS applications, SIEM, SOAR, Incident Response Tooling, scripting language skills"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_db7b0f51-7df"},"title":"Senior Cloud Support Engineer","description":"<p>As a Senior Cloud Support Engineer at CoreWeave, you&#39;ll be on the front lines of a technological revolution, empowering our customers to harness the full potential of our advanced Kubernetes-powered HPC cloud infrastructure.</p>\n<p>You&#39;ll be hands-on, collaborating with engineers and researchers to resolve issues that impact high-profile, mission-critical applications and cutting-edge AI training workloads. Your contributions will be pivotal in ensuring seamless performance, reliability, and success for our customers, positioning you at the very core of transformative technologies reshaping industries worldwide at a company that is truly one of a kind.</p>\n<p>In this role, you will:</p>\n<ul>\n<li>Guide and mentor team members in developing their technical skills and troubleshooting capabilities across all disciplines supported by CoreWeave.</li>\n<li>Provide real-time feedback and coaching, reviewing tickets to identify opportunities for improvement and ensure quality assurance (QA).</li>\n<li>Develop and deliver training sessions to improve the team&#39;s proficiency and efficiency in resolving customer issues.</li>\n<li>Use technical expertise to investigate, debug, and resolve customer-impacting issues with the curiosity required to uncover and understand root causes.</li>\n<li>Maintain high customer satisfaction through swift, accurate, and empathetic high-touch support communications, as well as established best practices.</li>\n<li>Help design and implement troubleshooting best practices to ensure fast, accurate client resolutions.</li>\n<li>Contribute to refining processes, workflows, and playbooks for handling complex customer challenges.</li>\n<li>Serve as a technical escalation point for high-priority escalations or complex cases, modeling effective problem-solving approaches.</li>\n<li>Lead the creation of knowledge-sharing resources, including documentation, tutorials, and how-to guides.</li>\n<li>Enhance the support team&#39;s knowledge of CoreWeave&#39;s products and services through continuous learning initiatives.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>Have a Bachelor&#39;s degree in Information Science / Information Technology, Data Science, Computer Science, Engineering, Mathematics, Physics, or a related field, OR equivalent experience in a technical position</li>\n<li>At least 5+ years of experience in cloud support, systems administration, or related technical support-focused roles</li>\n<li>Proven hands-on work experience with Kubernetes</li>\n<li>Experience with networking, load balancing, storage volumes, observability, node management, High-Performance Computing (HPC), and Linux system administration</li>\n<li>Proven ability to mentor team members, foster technical growth, and improve team-wide capabilities through guidance and feedback</li>\n<li>Experience with observability tools such as Grafana</li>\n<li>Strong troubleshooting skills, with experience resolving complex customer issues and driving quality assurance through ticket reviews or similar processes</li>\n<li>Demonstrated success collaborating with cross-functional teams to refine workflows, implement best practices, and advocate for necessary tools or process changes</li>\n<li>Excellent written and verbal communication skills, with a track record of simplifying complex concepts for diverse audiences</li>\n<li>Strong technical presentation skills, with experience delivering precise, engaging, and informative presentations to technical and non-technical audiences, effectively showcasing complex concepts and solutions</li>\n</ul>\n<p>Preferred:</p>\n<ul>\n<li>CKA Certified</li>\n<li>Demonstrated experience with training, coaching, and creating onboarding materials.</li>\n<li>Operates in a fast-paced, global, 24/7 support team environment</li>\n<li>Ability to collaborate across different time zones</li>\n<li>On-site office environment, hybrid, or remote options depending on location</li>\n<li>Flexible to travel up to 10% (~25 days/year)</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too.</p>\n<p>Come join us!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_db7b0f51-7df","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4568136006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$122,000 to $163,000","x-skills-required":["cloud support","systems administration","Kubernetes","networking","load balancing","storage volumes","observability","node management","High-Performance Computing (HPC)","Linux system administration"],"x-skills-preferred":["CKA Certified","training","coaching","onboarding materials","fast-paced global support team environment","collaboration across different time zones"],"datePosted":"2026-04-18T15:49:50.841Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"cloud support, systems administration, Kubernetes, networking, load balancing, storage volumes, observability, node management, High-Performance Computing (HPC), Linux system administration, CKA Certified, training, coaching, onboarding materials, fast-paced global support team environment, collaboration across different time zones","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":122000,"maxValue":163000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c571d7f7-d82"},"title":"Engineering Manager - Storage","description":"<p>At Databricks, we are building and running the world&#39;s best data and AI infrastructure platform. As an Engineering Manager, you will work with your team to build mission-critical Lakebase services on the Databricks Platform at scale.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Drive continuous delivery within a team of experts in storage technology, distributed systems and Rust.</li>\n<li>Manage the development and rollout of storage services that host millions of customer databases across dozens of regions</li>\n<li>Partner with peer engineering teams across Databricks to co-evolve Lakebase services with our global infrastructure.</li>\n<li>Lead operational excellence in 24/7 operation of our system</li>\n</ul>\n<p>The impact you will have:</p>\n<ul>\n<li>Hire great engineers to build an outstanding team.</li>\n<li>Support engineers in their career development by providing clear feedback and develop engineering leaders.</li>\n<li>Ensure high technical standards by instituting processes (architecture reviews, testing) and culture (engineering excellence).</li>\n<li>Work with engineering and product leadership to build a long-term roadmap.</li>\n<li>Coordinate execution and collaborate across teams to unblock cross-cutting projects.</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>Experience with building and shipping storage systems where correctness and performance are essential</li>\n<li>BS (or higher) in Computer Science, or a related field</li>\n<li>2+ years of experience building and leading a team of engineers working in a related system</li>\n<li>Experience with build, release and deployment infrastructure technologies such as Spinnaker, Jenkins, Airflow, Docker, Kubernetes, Terraform, Bazel, etc.</li>\n<li>Ability to attract, hire, and coach engineers who meet the Databricks hiring standards</li>\n<li>Comfort working on cross-functional projects with the ability to deeply understand product and customer personas</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c571d7f7-d82","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8476581002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["storage technology","distributed systems","Rust","Spinnaker","Jenkins","Airflow","Docker","Kubernetes","Terraform","Bazel"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:50.298Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"storage technology, distributed systems, Rust, Spinnaker, Jenkins, Airflow, Docker, Kubernetes, Terraform, Bazel"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6aab7ed8-23a"},"title":"Senior Software Engineer - Data","description":"<p>We are seeking an experienced Senior Software Engineer (Data) to join our fast-paced, collaborative data team. In this role, you will have broad authority to drive the direction of our technographic data services, building world-class data pipelines and systems to process billions of signals and data points.</p>\n<p>This is an exciting opportunity to solve challenging problems and make a big impact as we invest in making technographics a first-class offering.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Build and optimize big data pipelines to extract and process signals from the web, job postings, and other sources</li>\n<li>Design and implement data architectures and storage solutions to efficiently handle massive data volumes</li>\n<li>Collaborate closely with data scientists to support and integrate ML models into data workflows</li>\n<li>Continuously improve data quality, performance, and scalability of our technographic data platform</li>\n<li>Drive technical strategy and roadmap for the data processing infrastructure</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Extensive experience building and scaling big data pipelines and architectures from scratch</li>\n<li>Deep expertise in big data frameworks (Hadoop, Spark) and the JVM stack (Java, Scala)</li>\n<li>Strong software engineering fundamentals and ability to write efficient, high-quality code</li>\n<li>Experience with entity recognition and NLP techniques a plus</li>\n<li>Proven track record delivering results and driving projects in a fast-paced environment</li>\n<li>Excellent collaboration and communication skills to work with data scientists, analysts and product teams</li>\n<li>Passion for leveraging huge datasets to power valuable insights</li>\n</ul>\n<p>Ideal Background:</p>\n<ul>\n<li>8+ years of experience in software engineering roles</li>\n<li>Experience working with very large datasets and distributed systems</li>\n<li>Familiarity building data pipelines at large tech companies or data-driven organisations</li>\n<li>Bachelor&#39;s or advanced degree in Computer Science, Engineering or related technical field</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6aab7ed8-23a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8486808002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$140,000-$220,000 USD","x-skills-required":["big data pipelines","data architectures","storage solutions","ML models","data quality","performance","scalability","data processing infrastructure","Hadoop","Spark","Java","Scala","entity recognition","NLP techniques"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:24.766Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bethesda, Maryland, United States; Waltham, Massachusetts, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"big data pipelines, data architectures, storage solutions, ML models, data quality, performance, scalability, data processing infrastructure, Hadoop, Spark, Java, Scala, entity recognition, NLP techniques","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":140000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_70a6eadc-7c1"},"title":"Security Programs - Technical Program Manager","description":"<p>We are seeking a Security Technical Program Manager to join our Product Engineering organization. As a Security Technical Program Manager, you will work across cross-functional teams to ensure our cloud infrastructure is secure and private, while maintaining scalability and delivery of exceptional performance to meet the demands of our customers.</p>\n<p>The ideal candidate will have 8+ years of hands-on experience in Security Technical Program Management, Security Strategy, Security Risk Management and/or Security Compliance roles, ideally within the cloud services industry. They will have a Bachelor&#39;s degree in Information Security, Computer Science, or a related field or equivalent job experience.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Lead end-to-end program management for critical security engineering and security compliance initiatives, including cross-functional planning, execution, delivery, and retrospectives</li>\n<li>Define program scope, milestones, and success metrics while managing security risks and dependencies</li>\n<li>Partner closely within the security team, and across engineering, product management and operations teams to ensure alignment on priorities and deliverables</li>\n<li>Act as the primary point of contact for security and cross-functional stakeholders, providing regular status updates, addressing risks, and ensuring accountability</li>\n<li>Facilitate and influence technical security, privacy and compliance discussions and decisions to align with long-term infrastructure goals and business objectives</li>\n<li>Develop and implement scalable processes to improve efficiency and predictability in program delivery</li>\n<li>Strategically automate and improve day-to-day operations, processes and reporting</li>\n<li>Tailor communications to a diverse audience and remain adaptable to a wide range of personalities and technical depth</li>\n</ul>\n<p>What We Offer:</p>\n<ul>\n<li>Competitive salary range of $122,000 to $237,000</li>\n<li>Discretionary bonus, equity awards, and a comprehensive benefits program</li>\n<li>Medical, dental, and vision insurance - 100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace:</p>\n<ul>\n<li>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_70a6eadc-7c1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4556342006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$122,000 to $237,000","x-skills-required":["Security Technical Program Management","Security Strategy","Security Risk Management","Security Compliance","Cloud Services","Program Management","Cross-Functional Team Collaboration","Communication","Adaptability","Technical Security","Privacy","Compliance"],"x-skills-preferred":["Networking","Storage","Containerization (Kubernetes)","CI/CD Pipelines"],"datePosted":"2026-04-18T15:49:22.921Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Security Technical Program Management, Security Strategy, Security Risk Management, Security Compliance, Cloud Services, Program Management, Cross-Functional Team Collaboration, Communication, Adaptability, Technical Security, Privacy, Compliance, Networking, Storage, Containerization (Kubernetes), CI/CD Pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":122000,"maxValue":237000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9a2bbb70-2c0"},"title":"Senior Software Engineer - Data Platform","description":"<p>We are seeking a Senior Software Engineer to join our team in Bengaluru, India. As a Senior Software Engineer at Databricks, you will be responsible for designing, developing, and deploying large-scale distributed systems, including backend, DDS, and full-stack engineering. You will work closely with our product management team to bring great user experiences to our customers.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and develop reliable and high-performance services and client libraries for storing and accessing large amounts of data on cloud storage backends, such as AWS S3 and Azure Blob Store.</li>\n<li>Build scalable services using Scala, Kubernetes, and data pipelines, such as Apache Spark and Databricks.</li>\n<li>Work on a SaaS platform or with Service-Oriented Architectures.</li>\n<li>Collaborate with our DDS team to develop and deploy data-centric solutions using Apache Spark, Data Plane Storage, Delta Lake, and Delta Pipelines.</li>\n<li>Develop and maintain high-quality code, following best practices and coding standards.</li>\n<li>Participate in code reviews and provide feedback to improve the quality of the codebase.</li>\n<li>Troubleshoot and resolve issues that arise during deployment and operation.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science or a related field.</li>\n<li>7+ years of production-level experience in one of the following languages: Python, Java, Scala, C++, or similar language.</li>\n<li>Experience developing large-scale distributed systems from scratch.</li>\n<li>Experience working on a SaaS platform or with Service-Oriented Architectures.</li>\n<li>Strong understanding of software design patterns and principles.</li>\n<li>Excellent problem-solving skills and attention to detail.</li>\n<li>Ability to work effectively in a team environment.</li>\n<li>Strong communication and collaboration skills.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with Apache Spark, Data Plane Storage, Delta Lake, and Delta Pipelines.</li>\n<li>Knowledge of cloud-based storage systems, such as AWS S3 and Azure Blob Store.</li>\n<li>Familiarity with containerization using Docker and Kubernetes.</li>\n<li>Experience with continuous integration and continuous deployment (CI/CD) pipelines.</li>\n<li>Strong understanding of security principles and practices.</li>\n<li>Familiarity with agile development methodologies and version control systems, such as Git.</li>\n</ul>\n<p>Benefits:</p>\n<p>At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please click here.</p>\n<p>Our Commitment to Diversity and Inclusion:</p>\n<p>Databricks is an equal opportunities employer and welcomes applications from diverse candidates. We are committed to creating an inclusive and respectful work environment where everyone feels valued and empowered to contribute their best work.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9a2bbb70-2c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7601580002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","Scala","C++","Apache Spark","Data Plane Storage","Delta Lake","Delta Pipelines","Kubernetes","Docker","Git","Agile development methodologies","Version control systems"],"x-skills-preferred":["Cloud-based storage systems","Containerization","Continuous integration and continuous deployment (CI/CD) pipelines","Security principles and practices"],"datePosted":"2026-04-18T15:49:17.527Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, C++, Apache Spark, Data Plane Storage, Delta Lake, Delta Pipelines, Kubernetes, Docker, Git, Agile development methodologies, Version control systems, Cloud-based storage systems, Containerization, Continuous integration and continuous deployment (CI/CD) pipelines, Security principles and practices"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6de6ae31-894"},"title":"Powertrain Lead (Omen)","description":"<p>We are seeking a Powertrain Lead to join our Omen team at our Costa Mesa HQ. As the world enters an era of strategic competition, we are committed to bringing cutting-edge autonomy, AI, computer vision, sensor fusion, and networking technology to the military in months, not years.</p>\n<p>In this role, you will lead the design, development, and integration of Omen&#39;s advanced electric propulsion system. You will own the electric powertrain architecture from concept through production, optimizing for the unique demands of a tail-sitter platform including VTOL operations, efficient cruise flight, and charging.</p>\n<p>As the electric propulsion Lead, you will work closely with program leadership and cross-functional teammates to deliver a high-performance, reliable solution which seamlessly integrates with the rotor systems, powerplant, and cooling systems. You will own the propeller drive system as well as electric components of the on-board power generation, including battery, power management, and power conversion.</p>\n<p>You will drive component selection, manage supplier relationships, lead testing and validation efforts, and ensure the electric propulsion systems meet performance and integration requirements. This is a hands-on high ownership role where you will tackle hard technical problems, be the expert, and play a critical role in bringing a groundbreaking aircraft platform to operational reality.</p>\n<p>Join our dynamic team and help us build the next generation of autonomous flight systems.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Own the design of electric propulsion systems including motors, battery, inverters, and high voltage power management</li>\n<li>Lead electric performance analysis, voltage and thermal modeling, and energy optimization studies to meet mission requirements for electrical loads.</li>\n<li>Lead the development and execution of test plans for electric propulsion subsystems and combined power generation, including ground test stands, HILs, and flight test, building up to system verification to ensure compliance with project goals and industry standards</li>\n<li>Define and validate technical performance measures including power, efficiency, thermal limits, and fault tolerance across all operating modes</li>\n<li>Engage suppliers and create detailed technical specifications and interface control documents for electric propulsion components and assemblies as required</li>\n<li>Collaborate with other teams for solutions on system packaging, structural/thermal interfaces, and thrust/power control architectures</li>\n<li>Manage development schedules and technical risks in coordination with program management</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>High-ownership and a bias toward action – if you see a problem, you want to fix it</li>\n<li>Bachelor&#39;s degree in Aerospace Engineering, Mechanical Engineer, Robotics, Physics, or related technical discipline</li>\n<li>12+ professional years or be recognized as a seasoned expert with significant influence in their field</li>\n<li>Experience in engineering leadership and project execution for aerospace, automotive, or defense sectors with focus on electric or hybrid propulsion</li>\n<li>In-depth knowledge of electric motors, generators, power electronics, battery systems, thermal management, and energy storage architectures</li>\n<li>Hands-on experience with propulsion system development, testing, and integration for aerospace platforms</li>\n<li>Strong project management skills, with experience using tools like JIRA and Confluence, and knowledge of Agile and Waterfall methodologies</li>\n<li>Exceptional communication and stakeholder management abilities, capable of engaging effectively with both technical teams and non-technical partners</li>\n<li>Eligible to obtain and maintain an active U.S. Secret security clearance</li>\n<li>Travel to test sites and vendors as needed</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Direct experience developing hybrid-electric or electric propulsion systems for aircraft, rotorcraft, eVTOL, or UAS platforms</li>\n<li>Proven track record leading propulsion systems through full qualification campaigns from early development through flight test and production</li>\n<li>Experience with propulsion system modeling and simulation tools</li>\n<li>Familiarity with aviation environmental and EMI standards (MIL-STD-810, MIL-STD-461, DO-160, or equivalent) and qualification campaigns for flight hardware</li>\n<li>Experience with Formula SAE, Baja SAE, Design Build Fly, or similar competitive engineering programs demonstrating hands-on vehicle integration experience</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6de6ae31-894","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.andurilindustries.com/","logo":"https://logos.yubhub.co/andurilindustries.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5074209007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$250,000 USD","x-skills-required":["Aerospace Engineering","Mechanical Engineer","Robotics","Physics","Electric Motors","Generators","Power Electronics","Battery Systems","Thermal Management","Energy Storage Architectures","Propulsion System Development","Testing","Integration","Project Management","Agile","Waterfall","JIRA","Confluence"],"x-skills-preferred":["Hybrid-Electric Propulsion","Electric Propulsion","Propulsion System Modeling","Simulation Tools","Aviation Environmental Standards","EMI Standards","Flight Hardware Qualification","Formula SAE","Baja SAE","Design Build Fly"],"datePosted":"2026-04-18T15:49:11.354Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Aerospace Engineering, Mechanical Engineer, Robotics, Physics, Electric Motors, Generators, Power Electronics, Battery Systems, Thermal Management, Energy Storage Architectures, Propulsion System Development, Testing, Integration, Project Management, Agile, Waterfall, JIRA, Confluence, Hybrid-Electric Propulsion, Electric Propulsion, Propulsion System Modeling, Simulation Tools, Aviation Environmental Standards, EMI Standards, Flight Hardware Qualification, Formula SAE, Baja SAE, Design Build Fly","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2d198020-3d5"},"title":"Sr. Engineer, Storage","description":"<p>The Storage Engine Team at CoreWeave is responsible for the product capabilities and data plane function of CoreWeave&#39;s managed storage products. We build reliable, scalable storage solutions with segment leading performance. Storage engine works with engineering teams across infrastructure, compute, and platform to ensure our storage services meet the needs of the world&#39;s most demanding AI workloads.</p>\n<p>The primary responsibilities of this role include designing and implementing distributed storage solutions to support scaling data-intensive AI workloads, contributing to the development of exabyte-scale, S3-compatible object storage, and integrating dedicated storage clusters into diverse customer environments. Additionally, the successful candidate will work with technologies such as RDMA, GPU Direct Storage, and distributed filesystems protocols such as NFS or FUSE to optimize storage performance and efficiency.</p>\n<p>Key responsibilities also include leading efforts to improve the reliability, durability, security, and observability of our storage stack, collaborating with operations teams to monitor, troubleshoot, and improve storage systems in production environments, setting the bar for developing metrics and dashboards to provide visibility into storage performance and health, analyzing telemetry and system data to drive improvements in throughput, latency, and resilience, and working cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack.</p>\n<p>A key aspect of this role is sharing knowledge and mentoring other engineers on best practices in building distributed, high-performance systems.</p>\n<p>To be successful in this role, the ideal candidate will have a strong background in storage systems engineering or infrastructure, with a minimum of 8-10 years of experience. They will also have hands-on experience with object storage or distributed filesystems in production environments, as well as proficiency in a systems programming language such as Go, C, or Rust. Additionally, they will have experience working with cloud-native infrastructure, Kubernetes, and scalable system architectures, and familiarity with storage observability tools and telemetry pipelines.</p>\n<p>If you&#39;re a motivated and experienced engineer looking to join a dynamic team and contribute to the development of cutting-edge storage solutions, we encourage you to apply for this exciting opportunity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2d198020-3d5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4664429006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$143,000 to $210,000","x-skills-required":["storage systems engineering","infrastructure","object storage","distributed filesystems","RDMA","GPU Direct Storage","NFS","FUSE","cloud-native infrastructure","Kubernetes","scalable system architectures","storage observability tools","telemetry pipelines"],"x-skills-preferred":["Go","C","Rust","distributed systems","high-performance systems","storage performance and efficiency"],"datePosted":"2026-04-18T15:49:07.662Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"storage systems engineering, infrastructure, object storage, distributed filesystems, RDMA, GPU Direct Storage, NFS, FUSE, cloud-native infrastructure, Kubernetes, scalable system architectures, storage observability tools, telemetry pipelines, Go, C, Rust, distributed systems, high-performance systems, storage performance and efficiency","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":143000,"maxValue":210000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d7e1a365-9dd"},"title":"Principal Software Engineer II - Search Management - Elasticsearch","description":"<p>We&#39;re looking for a Principal Software Engineer to join the Elasticsearch - Search Management team. This globally-distributed team of experienced engineers focuses on delivering a robust and feature-rich search experience, including contributing to improving the search experience in Lucene.</p>\n<p>As a Principal Software Engineer, you will be a full-time Elasticsearch contributor, building data-intensive new features and fixing intriguing bugs, all while making the code easier to understand. You&#39;ll work with a globally distributed team of experienced engineers focused on the search capabilities of Elasticsearch.</p>\n<p>You&#39;ll be an expert in several areas of Elasticsearch and everyone will turn to you when they have a question about them. You&#39;ll improve those areas based on your questions and your instincts.</p>\n<p>You&#39;ll help us create the future of search within Elasticsearch - building a scalable search tier for our Serverless platform and writing search functionality in ES|QL, our new piped query language as two examples.</p>\n<p>You&#39;ll work with community members from all over the world on issues and pull requests, sometimes triaging them and handing them off to other experts and sometimes handling them yourself.</p>\n<p>You&#39;ll write idiomatic modern Java -- Elasticsearch is 99.8% Java!</p>\n<p>We&#39;re looking for someone with strong skills in core Java and a conversant in the standard library of data structures and concurrency constructs, as well as newer features like lambdas. You should be comfortable developing collaboratively, giving and receiving feedback on code and approaches and APIs.</p>\n<p>You&#39;ve used several data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra and have some idea how they work and why they work that way.</p>\n<p>You have excellent verbal and written communication skills. Like we said, collaborating on the internet is hard. We try to be respectful, empathetic, and trusting in all of our interactions. And we&#39;d expect that from you too.</p>\n<p>Bonus points if you&#39;ve built things with Elasticsearch before, worked in the search and information retrieval space, or have experience writing code for software-as-a-service or platforms-as-a-service.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d7e1a365-9dd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Elastic","sameAs":"https://www.elastic.co/","logo":"https://logos.yubhub.co/elastic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/elastic/jobs/7699084","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$154,000-$243,600 CAD","x-skills-required":["core Java","standard library of data structures and concurrency constructs","newer features like lambdas","data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra","idiomatic modern Java"],"x-skills-preferred":["search and information retrieval space","software-as-a-service or platforms-as-a-service","collaborative development","code review","API design"],"datePosted":"2026-04-18T15:49:07.452Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"core Java, standard library of data structures and concurrency constructs, newer features like lambdas, data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra, idiomatic modern Java, search and information retrieval space, software-as-a-service or platforms-as-a-service, collaborative development, code review, API design","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":154000,"maxValue":243600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a3e5c23b-397"},"title":"Sustainable Data Center Design Manager","description":"<p><strong>Job Title: Sustainable Data Center Design Manager</strong></p>\n<p>We are seeking a highly skilled and experienced Sustainable Data Center Design Manager to lead sustainability-driven design initiatives across our data center construction projects and engineering design.</p>\n<p><strong>About the Role:</strong></p>\n<p>The focus of this role is on embedding sustainability into data center design, ensuring that projects integrate onsite renewables, energy-efficiency strategies, low-carbon materials, water-conservation measures, and alternative refrigerants.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Act as the bridge between corporate sustainability goals and data center design execution, ensuring that high-level sustainability objectives translate into practical, technically sound design solutions.</li>\n</ul>\n<ul>\n<li>Collaboration &amp; Technical Leadership – Partner with MEP engineers, architects, sustainability consultants, and construction teams to integrate sustainable solutions into data center designs, ensuring alignment with engineering specifications and sustainability objectives.</li>\n</ul>\n<ul>\n<li>Energy Efficiency Design – Lead efforts to optimize energy efficiency, integrating technologies like AI-driven predictive modeling, automated energy management systems, and smart grid capabilities for operational savings.</li>\n</ul>\n<ul>\n<li>Onsite Renewables Evaluation – Evaluate and, where feasible, implement solar PV, wind energy, and battery storage systems, analyzing their technical, financial, and environmental impacts.</li>\n</ul>\n<ul>\n<li>Grid Interactivity &amp; Demand Response – Design data centers with a focus on energy resilience, including peak shaving, demand response, and microgrid capabilities that support the local grid.</li>\n</ul>\n<ul>\n<li>Lifecycle Carbon Analysis – Conduct carbon lifecycle assessments of design options, helping to identify and implement strategies that reduce overall greenhouse gas emissions.</li>\n</ul>\n<ul>\n<li>Water Conservation Design – Develop water-efficient systems, integrating renewable-powered cooling technologies and exploring innovative solutions like rainwater harvesting and closed-loop water systems.</li>\n</ul>\n<ul>\n<li>Sustainable Materials &amp; Construction – Recommend the use of sustainable, low-carbon building materials and ensure compliance with green building standards such as LEED.</li>\n</ul>\n<ul>\n<li>Alternative Refrigerants – Implement and advocate for low-GWP refrigerants and other sustainable HVAC solutions that reduce the environmental impact of data center cooling.</li>\n</ul>\n<ul>\n<li>AI &amp; Automation in Design – Utilize AI tools and automation technologies to optimize building designs for energy efficiency, water conservation, and operational flexibility.</li>\n</ul>\n<ul>\n<li>Sustainability Certifications – Ensure designs meet or exceed the requirements for certifications like LEED, WELL, ENERGY STAR, and other green building standards.</li>\n</ul>\n<ul>\n<li>Construction Oversight – Oversee the construction and commissioning of sustainable design elements, ensuring compliance with both design specifications and sustainability targets.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>Bachelor&#39;s degree or equivalent experience in Renewable Energy, Sustainability, or related field.</li>\n</ul>\n<ul>\n<li>5+ years in data center design, renewable energy integration, or energy infrastructure development.</li>\n</ul>\n<ul>\n<li>Demonstrated expertise and strong knowledge of sustainable energy technologies, including solar PV, wind, battery storage, microgrid integration, energy efficiency systems, water conservation, and grid interactivity.</li>\n</ul>\n<ul>\n<li>Experience in applying financial analysis to energy projects, including cost justification, Return on Investment (ROI) analysis, and knowledge of relevant incentive programs for renewable energy and sustainability initiatives.</li>\n</ul>\n<ul>\n<li>Proven ability to manage the implementation of sustainable solutions from the design phase through construction and commissioning.</li>\n</ul>\n<ul>\n<li>Ability to effectively communicate and present complex technical and financial analyses to diverse audiences, including executives, engineering teams, and regulatory agencies.</li>\n</ul>\n<p><strong>Why CoreWeave?</strong></p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p><strong>Salary Range:</strong></p>\n<p>The base salary range for this role is $134,000 to $179,000. The starting salary will be determined by job-related knowledge, skills, experience, and the market location.</p>\n<p><strong>Benefits:</strong></p>\n<p>In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n<li>100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Flexible Paid Time Off (PTO) policy</li>\n<li>Opportunities for professional growth and development</li>\n</ul>\n<p><strong>Category:</strong></p>\n<p>Engineering</p>\n<p><strong>Industry:</strong></p>\n<p>Technology</p>\n<p><strong>Required Skills:</strong></p>\n<ul>\n<li>Sustainable energy technologies</li>\n<li>Solar PV</li>\n<li>Wind energy</li>\n<li>Battery storage</li>\n<li>Microgrid integration</li>\n<li>Energy efficiency systems</li>\n<li>Water conservation</li>\n<li>Grid interactivity</li>\n</ul>\n<p><strong>Preferred Skills:</strong></p>\n<ul>\n<li>AI tools and automation technologies</li>\n<li>LEED certification</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a3e5c23b-397","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4661968006","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"The base salary range for this role is $134,000 to $179,000.","x-skills-required":["Sustainable energy technologies","Solar PV","Wind energy","Battery storage","Microgrid integration","Energy efficiency systems","Water conservation","Grid interactivity"],"x-skills-preferred":["AI tools and automation technologies","LEED certification"],"datePosted":"2026-04-18T15:48:49.569Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / San Francisco, CA / Bellevue, WA / Richmond, VA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Sustainable energy technologies, Solar PV, Wind energy, Battery storage, Microgrid integration, Energy efficiency systems, Water conservation, Grid interactivity, AI tools and automation technologies, LEED certification","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":134000,"maxValue":179000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dc0c258f-1f6"},"title":"Engineering Manager II, Enterprise AI Solutions","description":"<p>We are seeking a Business Savvy Engineering Manager to help define the future of Corporate IT&#39;s AI-based future at Pinterest. Working closely with cross-functional engineering teams and business leaders, you will lead a nimble team playing a pivotal role in scaling Corporate IT&#39;s engineering department.</p>\n<p>As an Engineering Manager, you will guide your team in designing and building the solutions that make our business partners&#39; jobs easier, faster, and more capable. You will grow and empower engineers while shaping how we build Pinterest&#39;s AI future.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Lead a team of employees and contractors focused on solving business problems using AI tools.</li>\n<li>Work closely with the existing software engineering teams to develop a seamless and low-friction client experience.</li>\n<li>Mentor junior engineers to help them grow and develop into the best that they can be.</li>\n<li>Motivate and lead your team to show up every day and do their best work.</li>\n<li>Collaborate with stakeholders and partner teams across the organization to architect data lake storage and metadata management technologies to unlock big data and ML/AI innovations.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>2+ years of experience leading and growing engineering teams, with a strong hands-on background in Python.</li>\n<li>7+ years of industry experience designing, building, and operating scalable, highly available backend systems, including owning production-grade infrastructure at scale.</li>\n<li>Proficiency in designing and delivering AI-based solutions that solve real-world business problems.</li>\n<li>Understanding of business unit challenges and problems, focused on Finance, Accounting, Legal, Sales, and Marketing.</li>\n<li>Experience with cloud infrastructure on AWS and containerized services using Docker and Kubernetes.</li>\n<li>Demonstrated technical leadership and people management experience, including setting team vision and long-term roadmap, mentoring and growing engineers across all levels, driving day-to-day execution and engineering alignment, and partnering cross-functionally to deliver complex, high-impact platform investments.</li>\n<li>Demonstrated ability to use AI to accelerate team execution, system design, and decision-making, paired with sound judgment in validating outputs, maintaining quality, and taking ownership of final outcomes.</li>\n<li>Build storage capabilities that efficiently support large-scale ML/AI workloads, including high-throughput data access, schema evolution, and large-scale column backfills.</li>\n<li>Demonstrated ability to use AI to improve speed and quality in your day-to-day workflow for relevant outputs.</li>\n<li>High integrity and ownership: you protect sensitive data, avoid over-reliance on AI, and remain accountable for final decisions and deliverables.</li>\n</ul>\n<p>In-Office Requirement Statement:</p>\n<ul>\n<li>We let the type of work you do guide the collaboration style. That means we&#39;re not always working in an office, but we continue to gather for key moments of collaboration and connection.</li>\n<li>This role will need to be in the office for in-person collaboration 1-2 times/quarter, and therefore can be situated anywhere in the country.</li>\n</ul>\n<p>Relocation Statement:</p>\n<ul>\n<li>This position is not eligible for relocation assistance.</li>\n</ul>\n<p>At Pinterest, we believe the workplace should be equitable, inclusive, and inspiring for every employee. In an effort to provide greater transparency, we are sharing the base salary range for this position. The position is also eligible for equity. Final salary is based on a number of factors including location, travel, relevant prior experience, or particular skills and expertise.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dc0c258f-1f6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/7494960","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$177,185-$364,795 USD","x-skills-required":["Python","AI","Cloud infrastructure","Containerized services","Docker","Kubernetes","Data lake storage","Metadata management","Big data","ML/AI innovations"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:48:29.379Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, US; Remote, US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, AI, Cloud infrastructure, Containerized services, Docker, Kubernetes, Data lake storage, Metadata management, Big data, ML/AI innovations","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":177185,"maxValue":364795,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_716ddbeb-2fa"},"title":"Battery Engineer","description":"<p>We are seeking an experienced Battery Engineer to join our team. As a Battery Engineer, you will be responsible for the development and testing of energy storage systems for our defence and aerospace applications.</p>\n<p>Your primary responsibilities will include:</p>\n<ul>\n<li>Driving end-to-end energy storage system (ESS) development on extremely accelerated timelines</li>\n<li>Collaborating with program teams to define actionable ESS requirements from user story/CONOPs</li>\n<li>Executing and/or delivering the following ESS development artifacts/steps:</li>\n</ul>\n<ul>\n<li>ESS requirements, battery cell selection and validation</li>\n<li>Hands on experience with ESS Assembly and test</li>\n<li>Battery management system (BMS) integration and validation</li>\n<li>Cell, module, and pack level validation plan. Support bring-up, test, debug, and vehicle integration</li>\n<li>Ramp plan from development units to production</li>\n<li>Sustain fielded ESSs with root cause analysis and data processing, drive countermeasures/design updates as needed</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Bachelor’s Degree in Electrical Engineering, Mechanical Engineering or fundamental Physical Sciences with applied application focus on energy storage technologies, with 5+ years of experience designing, testing, and troubleshooting energy storage systems.</li>\n<li>Competence with battery test equipment such as cell and pack cyclers, thermal chambers, battery health testers (EIS, ACIR, hi-pot, megger, etc)</li>\n<li>Knowledge of modern battery management system capabilities and requirements in HW and SW/FW</li>\n<li>Experience in battery characterization methods: mission profile validation, equivalent-circuit model characterization, cycle/calendar life assessment, safety features, etc.</li>\n<li>Understanding of UN38.3, NAVSEA S9310 and/or equivalent (UL, IEC) battery safety standards</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Graduate Degree in Electrical Engineering, Mechanical Engineering or fundamental Physical Sciences with a focus on energy storage technologies</li>\n<li>Understanding and familiarity with MIL-STD-882, MIL-STD-810, MIL-STD-704, MIL-STD-461, MIL-STD-464</li>\n<li>Demonstrated experience in delivering robust, high-performance, energy storage systems with substantial technical ownership/contributions</li>\n<li>Demonstrated experience in battery testing, modeling, controls, and BMS integration/operation</li>\n</ul>\n<p>Salary Range: $132,000 - $213,000 USD</p>\n<p>Benefits:</p>\n<ul>\n<li>Comprehensive medical, dental, and vision plans at little to no cost to you</li>\n<li>Generous time off: Highly competitive PTO plans with a holiday hiatus in December. Caregiver &amp; Wellness Leave is available to care for family members, bond with a new baby, or address your own medical needs</li>\n<li>Family Planning &amp; Parenting Support: Coverage for fertility treatments (e.g., IVF, preservation), adoption, and gestational carriers, along with resources to support you and your partner from planning to parenting</li>\n<li>Mental Health Resources: Access free mental health resources 24/7, including therapy and life coaching. Additional work-life services, such as legal and financial support, are also available</li>\n<li>Professional Development: Annual reimbursement for professional development</li>\n<li>Commuter Benefits: Company-funded commuter benefits based on your region</li>\n<li>Relocation Assistance: Available depending on role eligibility</li>\n<li>Retirement Savings Plan: Traditional 401(k), Roth, and after-tax (mega backdoor Roth) options</li>\n</ul>\n<p>Please note that the salary range and benefits may vary based on the location and other factors.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_716ddbeb-2fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/4845566007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$132,000 - $213,000 USD","x-skills-required":["Battery Test Equipment","Battery Management System","Energy Storage Systems","Electrical Engineering","Mechanical Engineering"],"x-skills-preferred":["MIL-STD-882","MIL-STD-810","MIL-STD-704","MIL-STD-461","MIL-STD-464"],"datePosted":"2026-04-18T15:48:24.248Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Battery Test Equipment, Battery Management System, Energy Storage Systems, Electrical Engineering, Mechanical Engineering, MIL-STD-882, MIL-STD-810, MIL-STD-704, MIL-STD-461, MIL-STD-464","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":132000,"maxValue":213000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0ff568ca-d59"},"title":"Senior Software Engineer - Data Infrastructure Services","description":"<p>CoreWeave is seeking a senior software engineer to join its Data Platforms Team. The ideal candidate will have experience in database and stream processing, and will be responsible for designing and implementing the platform to deliver data to teams with a focus on providing managed solutions through APIs.</p>\n<p>The successful candidate will participate in operations and scaling of relational data platforms, develop a stream processing architecture, and improve the performance, security, reliability, and scalability of our data platforms and related services. They will also establish guidelines, guardrails for data access and storage for stakeholder teams, and ensure compliance with standards for data protection regulation.</p>\n<p>In addition to technical skills, the ideal candidate will be able to grow, change, invest in their teammates, be invested-in, share their ideas, listen to others, be curious, have fun, and be themselves. CoreWeave values diversity and inclusion, and encourages candidates from all backgrounds to apply.</p>\n<p>Key responsibilities:</p>\n<ul>\n<li>Design and implement the platform to deliver data to teams with a focus on providing managed solutions through APIs</li>\n</ul>\n<ul>\n<li>Participate in operations and scaling of relational data platforms</li>\n</ul>\n<ul>\n<li>Develop a stream processing architecture</li>\n</ul>\n<ul>\n<li>Improve the performance, security, reliability, and scalability of our data platforms and related services</li>\n</ul>\n<ul>\n<li>Establish guidelines, guardrails for data access and storage for stakeholder teams</li>\n</ul>\n<ul>\n<li>Ensure compliance with standards for data protection regulation</li>\n</ul>\n<ul>\n<li>Grow, change, invest in your teammates, be invested-in, share your ideas, listen to others, be curious, have fun, and be yourself</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>5+ years of experience in a software or infrastructure engineering industry</li>\n</ul>\n<ul>\n<li>Experience operating services in production and at scale</li>\n</ul>\n<ul>\n<li>Familiarity with one of the distributed NewSQL datastores such as CockroachDB, TiDB, YDB, Yugabyte and/or stream processing tools such as NATS or Kafka</li>\n</ul>\n<ul>\n<li>Experience with designing and operating these systems at scale</li>\n</ul>\n<ul>\n<li>Familiarity with Kubernetes and have interest or comfortable with using it for event-driven and/or stateful orchestration</li>\n</ul>\n<ul>\n<li>Proficiency in Go/Python/Java and interested in contributing to open source</li>\n</ul>\n<p>ExperienceLevel: senior</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0ff568ca-d59","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4671479006","x-work-arrangement":"hybrid","x-experience-level":null,"x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["database and stream processing","API design and implementation","operational and scaling of relational data platforms","stream processing architecture","performance, security, reliability, and scalability of data platforms","data access and storage guidelines","data protection regulation compliance"],"x-skills-preferred":["Kubernetes","Go/Python/Java","open source contribution"],"datePosted":"2026-04-18T15:48:10.919Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database and stream processing, API design and implementation, operational and scaling of relational data platforms, stream processing architecture, performance, security, reliability, and scalability of data platforms, data access and storage guidelines, data protection regulation compliance, Kubernetes, Go/Python/Java, open source contribution","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2c175171-5fd"},"title":"Enterprise Account Executive","description":"<p>We&#39;re looking for a motivated and talented Account Executive to join our growing team. At CoreWeave, we&#39;re building the next generation public cloud for accelerated workloads. Your primary responsibilities will be to understand client goals and establish how CoreWeave Cloud can meet their needs, specifically within Machine Learning and Batch Processing use cases.</p>\n<p>In this role, you will be responsible for regularly hitting quotas for new business and growing revenue across existing accounts. This position requires creativity, attention to detail, value-based sales skills, and an insatiable desire to help clients succeed.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Manage a sales pipeline and forecast revenue effectively.</li>\n<li>Interact with prospects and clients via Zoom, telephone, and email with a high attention to customization and detail.</li>\n<li>Consistently achieve quarterly and annual revenue quotas.</li>\n<li>Diligently update and maintain clean pipeline data using Salesforce.com.</li>\n<li>Quickly learn the CoreWeave product(s), effectively communicate the value proposition, and handle objections in competitive situations.</li>\n<li>Develop a deep understanding of client goals, pain points, and timelines in order to communicate priorities across the CoreWeave organization.</li>\n<li>Develop a deep understanding of our target markets, and stay up to date on industry news.</li>\n<li>Travel up to 35% to meet with clients and prospects, attend industry events, and participate in on-site meetings.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>10+ years of experience in B2B sales and/or account management roles.</li>\n<li>Minimum of 7 years Direct Enterprise selling experience.</li>\n<li>Track record of exceeding sales quotas.</li>\n<li>Excellent communication skills, with the ability to clearly explain complex concepts.</li>\n<li>Strong understanding of value-based sales best practices and procedures.</li>\n<li>Proven ability to perform successfully in a fast-paced and flexible environment.</li>\n<li>Passion for developing account expansion strategies and the value of post-sales customer success.</li>\n<li>Excellent time management and communication skills.</li>\n<li>Desire to drive change and evangelize new technologies in a way that resonates with customers.</li>\n<li>Ability to thrive in a fast-paced environment experience is key, a mix of large company and startup experience.</li>\n<li>The ability to speak with customers and partners at both a business-level and technical-level.</li>\n<li>Knowledge of the storage industry specifically: NFS, Object Storage and Enterprise Computing architectures.</li>\n<li>Competitive attitude and strong work ethic with the ability to enthusiastically represent the company.</li>\n<li>Proficient with Salesforce.com and Microsoft Office.</li>\n</ul>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>The base salary range for this role is $160,000–$190,000 per year with uncapped commission and total On Target Earnings (OTE) of $320,000–$380,000. The starting salary will be determined by job-related knowledge, skills, experience, and the market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we&#39;ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location. In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n<li>100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2c175171-5fd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4647802006","x-work-arrangement":"hybrid","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":"$160,000–$190,000 per year with uncapped commission and total On Target Earnings (OTE) of $320,000–$380,000","x-skills-required":["B2B sales","Account management","Salesforce.com","Microsoft Office","Value-based sales","Communication skills","Time management","Customer success"],"x-skills-preferred":["Cloud computing","Machine learning","Batch processing","Storage industry","NFS","Object Storage","Enterprise Computing architectures"],"datePosted":"2026-04-18T15:48:04.435Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA / Sunnyvale, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Sales","industry":"Technology","skills":"B2B sales, Account management, Salesforce.com, Microsoft Office, Value-based sales, Communication skills, Time management, Customer success, Cloud computing, Machine learning, Batch processing, Storage industry, NFS, Object Storage, Enterprise Computing architectures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":380000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5a5a8459-f04"},"title":"Engineering Manager of Managers, Data Platform","description":"<p>Job Description:</p>\n<p><strong>Who we are</strong></p>\n<p>Stripe is a financial infrastructure platform for businesses. Millions of companies - from the world’s largest enterprises to the most ambitious startups - use Stripe to accept payments, grow their revenue, and accelerate new business opportunities.</p>\n<p><strong>About the team</strong></p>\n<p>The Big Data Infrastructure organization is a globally distributed team of approximately 40 engineers spread across Dublin, Bangalore, Seattle, and San Francisco. This team is the backbone of the company’s data ecosystem, responsible for building, scaling, and maintaining the highly reliable platforms that power data storage, orchestration, and processing at scale.</p>\n<p>As the Head of Big Data Infra, you will lead a global, ~40-person engineering organization responsible for the foundational data platforms that drive the business. Reporting directly to the Head of Compute, you will define the strategic vision and roadmap for the company&#39;s data lake, orchestration pipelines, and batch computing environments.</p>\n<p>The team&#39;s technical portfolio spans four core domains:</p>\n<ul>\n<li>Datalake (Storage): Managing scalable cloud storage and metadata layers, leveraging Amazon S3, Apache Iceberg (metastore and integrations), SAL, and Hive Metastore (HMS).</li>\n</ul>\n<ul>\n<li>Data Orchestration: Ensuring robust pipeline execution and scheduling using Apache Airflow.</li>\n</ul>\n<ul>\n<li>Batch Compute Infra (Data Store): Maintaining foundational data infrastructure and legacy systems, including Hadoop.</li>\n</ul>\n<ul>\n<li>Batch Compute Experience (Data Processing): Optimizing and delivering powerful data processing environments utilizing Apache Spark and Apache Celeborn.</li>\n</ul>\n<p><strong>What you’ll do</strong></p>\n<p>You will move beyond day-to-day management to act as an industry leader, effectively advocating for your organization&#39;s mission and impact. You will be expected to see problems others don&#39;t and rally people to independently create solutions.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Set Strategic Vision: Define the scope, vision, and goals for your organization with little or no guidance. You will anticipate industry trends to influence Stripe&#39;s long-range plans and set direction on a multi-year timeframe.</li>\n</ul>\n<ul>\n<li>Lead at Scale: Manage the achievement of and accountability for broad swaths of programs. You will establish wide-ranging and scaled processes, anticipating and removing roadblocks across multiple teams.</li>\n</ul>\n<ul>\n<li>Drive Operational Excellence: Instill a culture of rigorous thinking and meticulous craftsmanship. You will ensure your organization drives constant improvement in team processes and maintains high standards of operational rigor.</li>\n</ul>\n<ul>\n<li>Indirect Influence: Use indirect influence to steer other teams toward making the right decisions for Stripe. You will effectively communicate your team&#39;s plan and how it links to Stripe&#39;s company vision to cross-functional stakeholders.</li>\n</ul>\n<ul>\n<li>Obsess Over Talent: Proactively invest in the development of the organization and its people at all levels. You will recruit world-class talent and coach your direct reports,who are themselves managers - to elevate the skills of the leadership team.</li>\n</ul>\n<ul>\n<li>Stewardship &amp; Culture: Act as an ambassador and advocate for Stripe, modeling ownership for all other Stripes. You will actively work to increase Stripe&#39;s inclusivity and diversity and use our operating principles to guide decision-making.</li>\n</ul>\n<p><strong>Who you are</strong></p>\n<p>We’re looking for someone who meets the minimum requirements to be considered for the role. If you meet these requirements, you are encouraged to apply. The preferred qualifications are a bonus, not a requirement.</p>\n<p><strong>Minimum requirements</strong></p>\n<ul>\n<li>Bachelor’s degree or equivalent practical experience with minimum 5 years of experience with software development.</li>\n</ul>\n<ul>\n<li>Minimum 5 years of experience in a technical leadership role; overseeing strategic projects.</li>\n</ul>\n<ul>\n<li>Minimum 3 years of Manager of Managers experience (managing other engineering managers).</li>\n</ul>\n<ul>\n<li>Experience building diverse teams to tackle challenging technical problems.</li>\n</ul>\n<ul>\n<li>Ability to thrive in a collaborative environment involving different stakeholders and subject matter experts.</li>\n</ul>\n<p><strong>Preferred qualifications</strong></p>\n<ul>\n<li>Strategic Ambiguity: Proven ability to translate chaos into clarity and navigate complex, high-impact work where you must define your own scope.</li>\n</ul>\n<ul>\n<li>Infrastructure at Scale: Successfully shipped and operated critical infrastructure with significant responsibility over funds or critical data.</li>\n</ul>\n<ul>\n<li>Cross-Functional Influence: A track record of getting other teams on board with your vision to support execution in a way that benefits the broader company.</li>\n</ul>\n<ul>\n<li>Curiosity: You enjoy learning and diving into the nuts-and-bolts of how things work (e.g., global money movement rails, currency conversion, or inter-company flows).</li>\n</ul>\n<ul>\n<li>Humility and Adaptability: You are humble and self-aware, with a history of adapting your management approach across different environments and seeking feedback to grow as a leader.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5a5a8459-f04","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7747391","x-work-arrangement":"onsite","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Strategic vision","Technical leadership","Project management","Team management","Communication","Problem-solving","Infrastructure at scale","Cross-functional influence","Curiosity","Humility and adaptability"],"x-skills-preferred":["Apache Iceberg","Apache Airflow","Apache Spark","Apache Celeborn","Amazon S3","Hive Metastore","SAL","Cloud storage","Metadata layers","Data orchestration","Batch computing infrastructure","Legacy systems","Hadoop","Global money movement rails","Currency conversion","Inter-company flows"],"datePosted":"2026-04-18T15:47:47.234Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Strategic vision, Technical leadership, Project management, Team management, Communication, Problem-solving, Infrastructure at scale, Cross-functional influence, Curiosity, Humility and adaptability, Apache Iceberg, Apache Airflow, Apache Spark, Apache Celeborn, Amazon S3, Hive Metastore, SAL, Cloud storage, Metadata layers, Data orchestration, Batch computing infrastructure, Legacy systems, Hadoop, Global money movement rails, Currency conversion, Inter-company flows"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3a40dbfa-d00"},"title":"Staff Software Engineer, Non-Human Identity","description":"<p>Secure Every Identity, from AI to Human Identity is the key to unlocking the potential of AI. Okta secures AI by building the trusted, neutral infrastructure that enables organisations to safely embrace this new era.</p>\n<p>We are looking for builders and owners who operate with speed and urgency and execute with excellence. This is an opportunity to do career-defining work. We&#39;re all in on this mission. If you are too, let&#39;s talk.</p>\n<p>The Team</p>\n<p>The Okta Privileged Access Management (PAM) team is building the future of identity for machines, services, and applications. We are seeking a world-class Staff Engineer to help us architect and build the high-performance core of our non-human identity platform.</p>\n<p>Your work, in close collaboration with our principal engineers and architects, will be the foundation of our strategy for managing privileged access in the modern enterprise. If you are a systems programmer who thrives on influencing the design of high-performance, concurrent, and resilient security software, this is the role for you.</p>\n<p>What you’ll be doing</p>\n<ul>\n<li>Contribute to Core Architecture:</li>\n<li>Partner with principal engineers and architects to design and implement a low-latency, high-throughput secrets engine for non-human identities</li>\n<li>Solve for Massive Scale:</li>\n<li>Write highly concurrent, performance-critical code capable of handling millions of machine-to-machine authentication and authorization requests</li>\n<li>Shape Technical Strategy:</li>\n<li>Play a key role in defining the long-term technical roadmap for scalability and performance, ensuring our platform can meet the demands of the largest enterprises</li>\n<li>Mentor and Elevate:</li>\n<li>As a senior engineer on the team, you will work with junior engineers to help them advance their SDLC expertise.</li>\n<li>On-Call:</li>\n<li>Participate in the rotational on-call activities with SRE and product development team</li>\n</ul>\n<p>What you’ll bring to the role</p>\n<ul>\n<li>Required Experience:</li>\n<li>8+ years of professional software engineering experience, with a heavy focus on backend or systems-level development</li>\n<li>Bachelor’s or Master’s degree in Computer Science, or equivalent practical experience</li>\n<li>Core Technical Expertise:</li>\n<li>Deep, hands-on expertise in multi-platform Go development and building high-performance, concurrent applications</li>\n<li>Experience designing or operating distributed systems</li>\n<li>Experience with secure systems (authn/authz, encryption, TLS, token handling, PKI, CAs, diagnosing TLS issues)</li>\n<li>Deep expertise in distributed storage systems, with a focus on replication, backup, and restore, and data management. (Postgres, etc.)</li>\n<li>Direct experience designing, building, or contributing to a secrets management, service mesh, or machine identity platform</li>\n<li>Expert-level at ergonomic API design (gRPC/openAPI), and building for reliability at scale</li>\n<li>Deep knowledge of cloud-native infrastructure</li>\n<li>Key Attributes:</li>\n<li>You are driven by the challenge of optimizing systems for performance, latency, and throughput, with a proven ability to diagnose complex, multi-system issues</li>\n<li>You have a proven track record of making significant contributions to the architecture of complex, mission-critical systems</li>\n<li>You thrive in an environment where you can focus on deep technical problems</li>\n<li>Bonus Points:</li>\n<li>Experience at a leading Cybersecurity or Infrastructure-as-Code company</li>\n<li>Contributions to open-source projects in the identity, security, or infrastructure space</li>\n</ul>\n<p>And extra credit if you have experience in any of the following!</p>\n<ul>\n<li>Deep expertise in backend systems engineering</li>\n<li>Experience building and scaling beyond standard three-tier monolithic architectures, with a focus on modern distributed systems</li>\n<li>Have worked on projects with complex, established systems</li>\n<li>Possess significant, hands-on experience in a Linux/Unix environment</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3a40dbfa-d00","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7674829","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$194,000-$267,000 USD","x-skills-required":["Go development","Distributed systems","Secure systems","Distributed storage systems","Secrets management","Service mesh","Machine identity platform","Ergonomic API design","Cloud-native infrastructure"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:47:43.090Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go development, Distributed systems, Secure systems, Distributed storage systems, Secrets management, Service mesh, Machine identity platform, Ergonomic API design, Cloud-native infrastructure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":194000,"maxValue":267000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8d0f0628-c25"},"title":"Staff Android Engineer, Guest & Host","description":"<p>As a product-centric Staff Android Engineer on the Reputation Clients team at Airbnb, you&#39;ll play a critical role in partnering with our product and design leads to drive our Quality-focused initiatives.</p>\n<p>This involves working with other engineers to define, build, and shape the future of Airbnb&#39;s Listings and Quality related product flows, such as reviews, performance insights, and supply onboarding and management.</p>\n<p>The team is highly collaborative with its partners in design, product, data science, legal, and marketing to deliver features and experiences used by millions of people.</p>\n<p>We work closely with other engineering teams to ensure we build systems that are performant and configurable with architecture that&#39;s scalable and robust.</p>\n<p>A typical day involves working closely with senior leaders to define and drive critical product systems that enable the company&#39;s future vision, collaborating with cross-functional partners to design and deliver high-quality user experiences, and establishing a technical vision and leading efforts to architect, build, and launch reliable, performant, and scalable solutions for Quality-related products.</p>\n<p>Your expertise includes 10+ years of relevant experience in the tech industry, with 1+ years of experience in an Architect or Staff Engineer role, and a strong sense for product experience and business outcomes.</p>\n<p>You&#39;re also expected to have excellent engineering and coding skills, and experience working with GraphQL and REST APIs, solid understanding of client networking, performance, caching, client storage solutions, accessibility, and asynchronous programming.</p>\n<p>The role requires strong communication and collaboration skills with both technical and non-technical partners, and the ability to influence non-technical partners such as product managers, design, and the business to develop a unified approach to overall architecture and product/design capabilities.</p>\n<p>The position is US-Remote Eligible, and you must live in a state where Airbnb, Inc. has a registered entity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8d0f0628-c25","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7681159","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$204,000-$255,000 USD","x-skills-required":["Android engineering","Kotlin","Jetpack Compose","GraphQL","REST APIs","Client networking","Performance","Caching","Client storage solutions","Accessibility","Asynchronous programming"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:47:40.923Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Android engineering, Kotlin, Jetpack Compose, GraphQL, REST APIs, Client networking, Performance, Caching, Client storage solutions, Accessibility, Asynchronous programming","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":204000,"maxValue":255000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_35f3735f-43a"},"title":"Senior Software Engineer - Database Engine Internals","description":"<p>Our mission at Databricks is to simplify the data lifecycle from ingestion to ETL, BI, and ML/AI with a unified platform.</p>\n<p>To achieve this goal, we believe the data warehouse architecture as we know it today will be replaced by a new architectural pattern, Lakehouse, open platforms that unify data warehousing and advanced analytics.</p>\n<p>A critical part of realizing this vision is the next generation (decoupled) query engine and structured storage system that can outperform specialized data warehouses in relational query performance, yet retain the expressiveness and of general purpose systems such as Apache Spark™ to support diverse workloads ranging from ETL to data science.</p>\n<p>As part of this team, you will be working in one or more of the following areas to design and implement these next gen systems that leapfrog state-of-the-art:</p>\n<ul>\n<li>Query compilation and optimization</li>\n<li>Distributed query execution and scheduling</li>\n<li>Vectorized execution engine</li>\n<li>Data security</li>\n<li>Resource management</li>\n<li>Transaction coordination</li>\n<li>Efficient storage structures (encodings, indexes)</li>\n<li>Automatic physical data optimization</li>\n</ul>\n<p>We look for:</p>\n<ul>\n<li>A passion for database systems, storage systems, distributed systems, language design, or performance optimization</li>\n<li>Experience working towards a multi-year vision with incremental deliverables</li>\n<li>Motivated by delivering customer value and impact</li>\n<li>5+ years of experience working in a related system (preferred)</li>\n</ul>\n<p>Pay Range Transparency</p>\n<p>Databricks is committed to fair and equitable compensation practices. The pay range(s) for this role is listed below and represents the expected salary range for non-commissionable roles or on-target earnings for commissionable roles.</p>\n<p>Actual compensation packages are based on several factors that are unique to each candidate, including but not limited to job-related skills, depth of experience, relevant certifications and training, and specific work location.</p>\n<p>Based on the factors above, Databricks anticipates utilizing the full width of the range.</p>\n<p>The total compensation package for this position may also include eligibility for annual performance bonus, equity, and the benefits listed above.</p>\n<p>For more information regarding which range your location is in visit our page here.</p>\n<p>Local Pay Range $166,000-$225,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_35f3735f-43a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/6544383002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$225,000 USD","x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimization","query compilation and optimization","distributed query execution and scheduling","vectorized execution engine","data security","resource management","transaction coordination","efficient storage structures"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:47:27.830Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimization, query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":225000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e1d2b108-713"},"title":"Oracle Fusion Software Developer","description":"<p>We are looking for an expert Oracle Integration Developer to join our Arsenal (Enterprise Systems) team. Your immediate mission: take ownership of our critical enterprise integrations connecting Oracle Fusion ERP with our upstream and downstream systems. These integrations, built on Oracle Integration Cloud, form the digital backbone that governs how we manage our business operations, from product data and procurement to manufacturing and financial processes.</p>\n<p>You will be tasked with stabilizing, optimizing, and making them exceptionally robust. Long-term, you will be the subject matter expert responsible for architecting and scaling our enterprise integration landscape. This is a high-impact role for someone who thrives on solving complex data challenges and wants to build the operational foundation that enables Anduril to scale its mission.</p>\n<p>The successful candidate will have 5+ years of hands-on experience developing complex integrations with deep specialization in Oracle Integration Cloud (OIC), specifically Oracle Integration 3. They will have proven experience integrating Oracle Fusion Cloud ERP with upstream and downstream enterprise systems, including deep familiarity with ERP data objects such as Items, BOMs, Suppliers, Purchase Orders, Work Orders, Inventory Transactions, and Financial data.</p>\n<p>Key responsibilities will include stabilizing and optimizing existing Oracle Fusion ERP integrations, architecting and building new enterprise integrations using Oracle Integration Cloud, owning the integration lifecycle, ensuring data integrity, collaborating and influencing with cross-functional teams, and leveraging modern Oracle Cloud tools.</p>\n<p>The ideal candidate will have excellent SQL skills, a strong command of XSLT, XPath, and complex data mapping, demonstrable experience building, securing, and consuming RESTful APIs and SOAP web services, and experience with Oracle Fusion ERP modules such as SCM (Supply Chain Management), Manufacturing, Procurement, or Financials.</p>\n<p>A tenacious problem-solver with a track record of troubleshooting, debugging, and stabilizing complex, business-critical systems, the successful candidate will be highly motivated, with a passion for delivering high-quality solutions and a commitment to continuous learning and improvement.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e1d2b108-713","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5058269007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$129,000-$171,000 USD","x-skills-required":["Oracle Integration Cloud","Oracle Fusion Cloud ERP","XSLT","XPath","RESTful APIs","SOAP web services","SQL","Oracle Fusion ERP modules (SCM, Manufacturing, Procurement, or Financials)"],"x-skills-preferred":["Oracle Visual Builder Cloud Service","Oracle Business Intelligence Cloud Connector","Oracle Cloud Infrastructure services (Functions, API Gateway, Object Storage, Logging, Autonomous Database)","PLM systems (Teamcenter, Windchill, Arena)","Git-based source control and CI/CD pipelines","Discrete manufacturing environment","Other programming languages (Python, Groovy, Java)"],"datePosted":"2026-04-18T15:47:23.789Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, Washington, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Oracle Integration Cloud, Oracle Fusion Cloud ERP, XSLT, XPath, RESTful APIs, SOAP web services, SQL, Oracle Fusion ERP modules (SCM, Manufacturing, Procurement, or Financials), Oracle Visual Builder Cloud Service, Oracle Business Intelligence Cloud Connector, Oracle Cloud Infrastructure services (Functions, API Gateway, Object Storage, Logging, Autonomous Database), PLM systems (Teamcenter, Windchill, Arena), Git-based source control and CI/CD pipelines, Discrete manufacturing environment, Other programming languages (Python, Groovy, Java)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":129000,"maxValue":171000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7fef2a53-3e7"},"title":"Staff Software Engineer, Storage","description":"<p>We&#39;re seeking a Staff Software Engineer to join our BE Platform team. As a Staff Software Engineer, you will play a key role in building and evolving control and data planes, improving underlying systems, and writing software that implements critical workflows to automate and enhance the operation of our large-scale storage infrastructure.</p>\n<p>Your responsibilities will include designing, writing, and delivering software to improve the availability, scalability, latency, and efficiency of Reddit&#39;s products in Go, C++, and sometimes Python. You will dive deep into the codebase of supported storage systems to understand system internals, make system-level improvements, enhancements, and implement complex code modifications. You will engage actively with the open-source community to implement and upstream changes to the OSS codebase.</p>\n<p>You will contribute to the design and implementation of high-performance, large-scale distributed storage systems to power various use cases at Reddit. You will collaborate closely with engineering teams and stakeholders to integrate storage capabilities into broader storage infrastructure and use cases across Reddit. You will mentor and guide other engineers on how to design, build, and evangelize vector storage services across Reddit.</p>\n<p>The ideal candidate will have 7+ years of experience building internet-scale software, preferably with a focus on machine learning storage infrastructure. They will have software development experience in one or more general-purpose programming languages, such as Golang, Python, C++, or Java. They will have hands-on experience implementing features, optimizations, and bug fixes to distributed storage systems. They will have experience contributing code improvements, features, and bug fixes to open-source projects.</p>\n<p>Additionally, the ideal candidate will have excellent communication skills to collaborate with a service-oriented team and company. They will be able to work effectively in a fast-paced environment and prioritize tasks to meet deadlines.</p>\n<p>We offer a comprehensive benefits package, including comprehensive healthcare benefits, income replacement programs, 401(k) match, family planning support, gender-affirming care, mental health and coaching benefits, flexible vacation, Reddit global days off, generous paid parental leave, and paid volunteer time off.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7fef2a53-3e7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7511679","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$217,000 - $303,900 USD","x-skills-required":["Go","C++","Python","mongoose","Golang","Java","Distributed storage systems","High-performance computing","Large-scale systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:47:00.406Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, C++, Python, mongoose, Golang, Java, Distributed storage systems, High-performance computing, Large-scale systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":217000,"maxValue":303900,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6af2312-b49"},"title":"Systems PhD - Software Engineer","description":"<p>We&#39;re looking for a Systems PhD - Software Engineer to join our Database Engine team. As a member of this team, you&#39;ll have opportunities to design and implement in many areas that leapfrog existing state-of-the-art systems, including query compilation &amp; optimisation, distributed query execution and scheduling, vectorised engine execution, data security, resource management, transaction coordination, efficient storage structures, and automatic physical data optimisation.</p>\n<p>To be successful in this role, you&#39;ll need a PhD in databases or systems, a passion for database systems, storage systems, distributed systems, language design, and/or performance optimisation, and be motivated by delivering customer value and impact.</p>\n<p>The pay range for this role is $140,000-$180,000 USD, and the total compensation package may also include eligibility for annual performance bonus, equity, and benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6af2312-b49","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8482086002","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$140,000-$180,000 USD","x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimisation","query compilation & optimisation","distributed query execution and scheduling","vectorised engine execution","data security","resource management","transaction coordination","efficient storage structures","automatic physical data optimisation"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:52.787Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bellevue, Washington; Seattle, Washington"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimisation, query compilation & optimisation, distributed query execution and scheduling, vectorised engine execution, data security, resource management, transaction coordination, efficient storage structures, automatic physical data optimisation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":140000,"maxValue":180000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_711f5c89-ed8"},"title":"Senior Staff Machine Learning Engineer, GenAI Platform","description":"<p>As a Senior Staff Machine Learning Engineer, you will help define and lead the vision for Reddit&#39;s large-scale GenAI Platform, shaping the strategy, architecture, and operating model that enable teams across the company to build, deploy, and scale generative AI products with confidence.</p>\n<p>Contribute to the design, implementation, and maintenance of the LLM Gateway, focusing on features like unified API endpoints for internal/externally hosted LLM, rate/token limit management, and intelligent failover mechanisms to boost uptime and reliability.</p>\n<p>Lead and execute the vision, strategy, and roadmap for Reddit&#39;s large-scale GenAI Platform.</p>\n<p>Define the platform architecture and operating model that enable teams to build, deploy, and scale GenAI products reliably.</p>\n<p>Drive the strategy for a unified LAG Gateway supporting internally and externally hosted LLMs through consistent APIs and abstractions.</p>\n<p>Set the direction for core platform capabilities such as rate and token limit management, intelligent failover, and production resilience.</p>\n<p>Shape Reddit&#39;s approach to an enterprise-grade RAG system.</p>\n<p>Establish the strategic direction for agentic AI workflows and tool-use patterns across the platform.</p>\n<p>Own the end-to-end platform strategy from concept through production adoption and long-term evolution.</p>\n<p>Drive MLOps and LLMOps standards across CI/CD, testing, versioning, evaluation, and lifecycle management.</p>\n<p>Define best practices for observability, monitoring, governance, and operational excellence across GenAI systems.</p>\n<p>Partner across engineering, product, and leadership to align platform investments with company priorities and user needs.</p>\n<p>Champion platform thinking with a strong focus on scalability, reliability, performance, and developer experience.</p>\n<p>Influence technical direction across teams by turning emerging AI capabilities into a scalable platform strategy.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_711f5c89-ed8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7772274","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$292,500-$409,500 USD","x-skills-required":["Machine Learning","GenAI Platform","LLM Gateway","API Endpoints","Rate/Token Limit Management","Intelligent Failover","Kubernetes","Cloud-Based Technologies","AWS","Google Cloud Storage","Infrastructure-as-Code","Terraform","Go","Python","CI/CD","Testing","Versioning","Evaluation","Lifecycle Management","Observability","Monitoring","Governance","Operational Excellence"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:48.652Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, GenAI Platform, LLM Gateway, API Endpoints, Rate/Token Limit Management, Intelligent Failover, Kubernetes, Cloud-Based Technologies, AWS, Google Cloud Storage, Infrastructure-as-Code, Terraform, Go, Python, CI/CD, Testing, Versioning, Evaluation, Lifecycle Management, Observability, Monitoring, Governance, Operational Excellence","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":292500,"maxValue":409500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8a144188-686"},"title":"Solutions Engineer, Benelux","description":"<p>At Cloudflare, we&#39;re on a mission to help build a better Internet. As a Solutions Engineer, you will be part of the Pre-Sales Solution Engineering organisation, owning the technical sale of the Cloudflare solution portfolio. You will work closely with our customers and partners to educate, empower, and ensure their success delivering Cloudflare security, reliability, and performance solutions.</p>\n<p>Your role will be to build champions and enable technical teams alongside our Benelux sales organisation to drive pipeline and close deals. As the technical advocate inside Cloudflare, you will work closely with teams across Sales, Product, Engineering, Customer Support and our channel partners to ensure our customers succeed with Cloudflare security, reliability, and performance solutions.</p>\n<p>We are looking for someone with strong experience in pre-sales, partner and account management, and excellent verbal and written communication skills in Dutch and English, suited for both technical and executive-level engagement. You should be comfortable speaking about the Cloudflare vision with all audiences.</p>\n<p>Specifically, we are looking for you to:</p>\n<ul>\n<li>Build and maintain long-term technical relationships with prospects, customers and ecosystem organisations across Benelux through demonstrating value, enablement, and uncovering new areas of potential revenue</li>\n<li>Drive technical solution design conversations through use case qualification and collaborative technical wins through demonstrations and proofs-of-concept</li>\n<li>Develop passionate technical champions within the technology ranks of your accounts, helping them drive sales for identified opportunities and build revenue pipeline</li>\n<li>Evangelize and represent Cloudflare through technical thought leadership and expertise</li>\n<li>Be the voice of the market internally at Cloudflare, engaging with and influencing Product and Engineering teams to meet the needs of your accounts and their customers</li>\n</ul>\n<p>You will travel requirement in the Benelux to support engagements, attend conferences and industry events, and collaborate with Cloudflare teammates.</p>\n<p>Examples of desirable skills, knowledge and experience include:</p>\n<ul>\n<li>Fluency in Dutch and English (verbal and written)</li>\n<li>Ability to communicate complex technical concepts to both technical and non-technical audiences, including C-level stakeholders</li>\n<li>Strong presentation and storytelling skills (whiteboarding, demos, executive briefings)</li>\n<li>Experience managing technical sales cycles end-to-end</li>\n<li>Ability to articulate business value and ROI of technical solutions, not just features</li>\n<li>Experience working within an integrated account team (alongside Account Executives, Customer Success, BDRs, and channel partners)</li>\n<li>Networking technologies including TCP/IP, UDP, DNS (authoritative and recursive, DNSSEC), IPv4/IPv6, BGP routing, Autonomous Systems, subnetting</li>\n<li>Tunneling and connectivity: GRE, IPsec, MPLS, SDWAN</li>\n<li>Cloud networking concepts: VPCs, peering, interconnect</li>\n<li>DDoS attack types (L3/L4/L7) and mitigation strategies</li>\n<li>Web Application Firewall (WAF) rule configuration and tuning</li>\n<li>VPN concepts and their limitations relative to Zero Trust approaches</li>\n<li>API security: API Gateway, rate limiting, schema validation, abuse prevention</li>\n<li>Bot management concepts and detection techniques</li>\n<li>SASE concepts and Zero Trust Networking architectures (ZTNA, CASB, SWG, DLP, RBI as integrated platform)</li>\n<li>Zero Trust Network Access (ZTNA) vs. traditional VPN architecture</li>\n<li>HTTP technologies and reverse proxy architecture: WAF, CDN, caching mechanics</li>\n<li>Detailed understanding of the flow from user to application, including hybrid cloud architectures</li>\n<li>Working knowledge of major cloud platforms: AWS, Azure, GCP (architecture patterns, native security tooling, VPC/peering models)</li>\n<li>Familiarity with Infrastructure-as-Code concepts (e.g. Terraform)</li>\n<li>Cloudflare Workers and the edge compute model (JavaScript/TypeScript)</li>\n<li>Familiarity with related primitives: KV, Object storage, serverless compute</li>\n<li>Familiarity with the competitive landscape across Cloudflare&#39;s product areas</li>\n<li>Understanding of why customers move from on-premises appliances to cloud-delivered security</li>\n<li>Awareness of relevant industry verticals: Financial Services, eCommerce, Gaming, Media, SaaS, Healthcare</li>\n</ul>\n<p>We value intellectual curiosity, adaptability, and a collaborative spirit. On the Solutions Engineering team, you will find an environment where everyone brings different strengths and jumps in to help each other. If you are passionate about technology and look forward to helping customers and ecosystem organisations realise the full promise of Cloudflare, we&#39;d love to hear from you.</p>\n<p>What makes Cloudflare special? We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet. Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8a144188-686","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7742347","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Networking technologies including TCP/IP, UDP, DNS (authoritative and recursive, DNSSEC), IPv4/IPv6, BGP routing, Autonomous Systems, subnetting","Tunneling and connectivity: GRE, IPsec, MPLS, SDWAN","Cloud networking concepts: VPCs, peering, interconnect","DDoS attack types (L3/L4/L7) and mitigation strategies","Web Application Firewall (WAF) rule configuration and tuning","VPN concepts and their limitations relative to Zero Trust approaches","API security: API Gateway, rate limiting, schema validation, abuse prevention","Bot management concepts and detection techniques","SASE concepts and Zero Trust Networking architectures (ZTNA, CASB, SWG, DLP, RBI as integrated platform)","Zero Trust Network Access (ZTNA) vs. traditional VPN architecture","HTTP technologies and reverse proxy architecture: WAF, CDN, caching mechanics","Detailed understanding of the flow from user to application, including hybrid cloud architectures","Working knowledge of major cloud platforms: AWS, Azure, GCP (architecture patterns, native security tooling, VPC/peering models)","Familiarity with Infrastructure-as-Code concepts (e.g. Terraform)","Cloudflare Workers and the edge compute model (JavaScript/TypeScript)","Familiarity with related primitives: KV, Object storage, serverless compute","Familiarity with the competitive landscape across Cloudflare's product areas","Understanding of why customers move from on-premises appliances to cloud-delivered security","Awareness of relevant industry verticals: Financial Services, eCommerce, Gaming, Media, SaaS, Healthcare"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:26.177Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Networking technologies including TCP/IP, UDP, DNS (authoritative and recursive, DNSSEC), IPv4/IPv6, BGP routing, Autonomous Systems, subnetting, Tunneling and connectivity: GRE, IPsec, MPLS, SDWAN, Cloud networking concepts: VPCs, peering, interconnect, DDoS attack types (L3/L4/L7) and mitigation strategies, Web Application Firewall (WAF) rule configuration and tuning, VPN concepts and their limitations relative to Zero Trust approaches, API security: API Gateway, rate limiting, schema validation, abuse prevention, Bot management concepts and detection techniques, SASE concepts and Zero Trust Networking architectures (ZTNA, CASB, SWG, DLP, RBI as integrated platform), Zero Trust Network Access (ZTNA) vs. traditional VPN architecture, HTTP technologies and reverse proxy architecture: WAF, CDN, caching mechanics, Detailed understanding of the flow from user to application, including hybrid cloud architectures, Working knowledge of major cloud platforms: AWS, Azure, GCP (architecture patterns, native security tooling, VPC/peering models), Familiarity with Infrastructure-as-Code concepts (e.g. Terraform), Cloudflare Workers and the edge compute model (JavaScript/TypeScript), Familiarity with related primitives: KV, Object storage, serverless compute, Familiarity with the competitive landscape across Cloudflare's product areas, Understanding of why customers move from on-premises appliances to cloud-delivered security, Awareness of relevant industry verticals: Financial Services, eCommerce, Gaming, Media, SaaS, Healthcare"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4f71a295-9c1"},"title":"Staff Software Engineer","description":"<p>We are looking for an experienced Staff Software Engineer to work on our next-generation Imports Platform team. The Imports Platform team is leading a strategic initiative to modernize Okta&#39;s identity lifecycle management capabilities by architecting and migrating from a legacy monolithic system to a highly scalable, distributed microservices platform.</p>\n<p>As a Staff Software Engineer on the Imports Platform team, you will be a technical leader who independently owns projects end-to-end, from ideation and architectural design through implementation, deployment, and operational excellence. You will drive technical strategy, make critical architectural decisions, and influence both your immediate team and cross-team initiatives.</p>\n<p>You will work on complex distributed systems challenges including massive-scale batch processing, real-time synchronization, and user matching algorithms that serve thousands of enterprise customers. This role requires strong technical leadership, strategic thinking, and the ability to balance short-term delivery with long-term platform health.</p>\n<p>You will mentor senior and junior engineers, partner with Product Management on feature strategy, and help shape the future of Okta&#39;s Imports platform. This is a hybrid position requiring a blend of remote and in-office collaboration.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Provide technical leadership on major development projects, including architectural design and implementation strategy</li>\n<li>Independently own and deliver projects end-to-end within the team, including technical prioritization and tradeoffs</li>\n<li>Generate design ideas and solutions for ambiguous problems, taking complete ownership from conception through production impact</li>\n<li>Design and architect core, high-performance, scalable software components with full ownership of all production aspects (scalability, reliability, monitoring, alerting, resource efficiency, testing, documentation)</li>\n<li>Lead technical design discussions and guide the team in making architectural decisions</li>\n<li>Drive the migration strategy from monolithic to microservices architecture, including planning, scoping, and execution</li>\n<li>Interface extensively with cross-functional teams (Architects, QA, Product, Technical Support, Documentation, UX, and SRE) to deliver comprehensive import and sync solutions</li>\n<li>Analyze and refine requirements with Product Management, partnering on product features and helping define the &#39;how&#39;</li>\n<li>Conduct code reviews with focus on systems design, reliability, performance, scalability, security, and maintainability</li>\n<li>Share knowledge widely, coordinate across teams, and manage risk and dependencies for projects</li>\n<li>Work with QA and SRE teams to define comprehensive testing strategies and operational excellence practices</li>\n<li>Independently troubleshoot complex production incidents spanning the home team, perform root cause analysis, and drive operational improvement projects</li>\n<li>Use data and metrics to drive technical decisions and validate the impact of architectural changes</li>\n<li>Mentor and provide technical guidance to senior and junior engineers on the team</li>\n<li>Help resolve difficult customer issues and work closely with Field teams and CSMs to identify patterns and drive product improvements</li>\n<li>Participate in group strategy discussions and help break down strategic initiatives into actionable technical milestones</li>\n<li>Proactively identify and advocate for improvements in team velocity, engineering practices, and operational processes</li>\n<li>Drive improvements in observability, monitoring, and production support capabilities</li>\n</ul>\n<p>Required Knowledge, Skills, and Abilities:</p>\n<ul>\n<li>7+ years of software development experience building highly-reliable, mission-critical software at scale</li>\n<li>Deep expertise with object-oriented languages, particularly Java, with proven ability to architect large-scale systems</li>\n<li>Expert-level knowledge of Spring Boot framework, Maven, and modern Java development practices</li>\n<li>Deep understanding of infrastructure-level technologies: distributed systems, caching strategies, stream processing, resilient architectures</li>\n<li>Solid experience with data stores including relational databases (MySQL), caching layers (Redis), and cloud storage (S3)</li>\n<li>Experience with one or more Directory services: Active Directory, LDAP, Office 365, Azure AD</li>\n<li>Strong experience with RESTful APIs, gRPC, and microservices architecture patterns</li>\n<li>Proven track record of working with systems at massive scale, including batch processing and real-time sync capabilities</li>\n<li>Experience with cloud platforms (AWS, GCP) including services like SQS, S3, and multi-region architectures</li>\n<li>Strong understanding of distributed job processing, message queues, and event-driven architectures</li>\n<li>Demonstrated ability to lead technical projects independently and influence cross-team initiatives</li>\n<li>Excellent communication skills with ability to share information widely and coordinate across teams</li>\n<li>Strong mentorship capabilities with experience guiding senior and junior engineers</li>\n<li>Customer-focused mindset with experience working with Field teams to resolve complex issues</li>\n<li>Strategic thinking ability to participate in and contribute to platform strategy</li>\n<li>Experience with operational excellence including incident management, root cause analysis, and driving systemic improvements</li>\n</ul>\n<p>Nice to Haves:</p>\n<ul>\n<li>Experience with Protocol Buffers (Protos) and building event-driven systems</li>\n<li>Experience with server-side technologies including advanced caching, asynchronous processing, multi-threading, and concurrency patterns</li>\n<li>Experience in Test-Driven Development (TDD) and automated testing strategies</li>\n<li>Deep knowledge of Identity and Access Management protocols and technologies: OAuth, OpenID Connect, SAML, SCIM, LDAP</li>\n<li>Experience with Microsoft Azure management APIs, Microsoft Graph API, Office 365, or ADFS</li>\n<li>Experience automating and deploying large-scale production services in AWS, GCP, or similar cloud platforms</li>\n<li>Experience with feature flag frameworks and gradual rollout strategies for large-scale migrations</li>\n<li>Understanding of user matching, correlation algorithms, and identity resolution at scale</li>\n<li>Experience with observability platforms, creating comprehensive monitoring and alerting strategies</li>\n<li>Experience migrating monolithic applications to microservices architecture</li>\n<li>Knowledge of data modeling for graph databases and relationship management</li>\n<li>Experience with incremental sync, delta detection, and change data capture patterns</li>\n<li>Background in building resilient systems with retry logic, circuit breakers, and failure handling</li>\n<li>Experience with performance optimization and capacity planning for high-throughput systems</li>\n</ul>\n<p>Education and Training:</p>\n<p>B.S. Computer Science or related field</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4f71a295-9c1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7725948","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Spring Boot","Maven","distributed systems","caching strategies","stream processing","resilient architectures","relational databases","caching layers","cloud storage","Directory services","RESTful APIs","gRPC","microservices architecture patterns","batch processing","real-time sync capabilities","cloud platforms","distributed job processing","message queues","event-driven architectures"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:19.715Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Spring Boot, Maven, distributed systems, caching strategies, stream processing, resilient architectures, relational databases, caching layers, cloud storage, Directory services, RESTful APIs, gRPC, microservices architecture patterns, batch processing, real-time sync capabilities, cloud platforms, distributed job processing, message queues, event-driven architectures"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8d587b18-3f1"},"title":"Staff Software Engineer - Database Engine Internals","description":"<p>We are seeking a Staff Software Engineer to join our team in Belgrade, Serbia. As a Staff Software Engineer, you will be part of a multi-year journey to design and build the next-generation query engine and structured storage system.</p>\n<p>Our mission at Databricks is to simplify the data lifecycle from ingestion to ETL, BI, and ML/AI with a unified platform. We believe the data warehouse architecture will be replaced by a new architectural pattern, Lakehouse, which unifies data warehousing and advanced analytics.</p>\n<p>The new architecture will help address several major challenges, including data staleness, reliability, total cost of ownership, data lock-in, and limited use-case support. To achieve this vision, we are building a decoupled query engine and structured storage system that can outperform specialized data warehouses in relational query performance.</p>\n<p>As a Staff Software Engineer, you will design these next-generation systems that leapfrog state-of-the-art within the following areas:</p>\n<ul>\n<li>Query compilation and optimization</li>\n<li>Distributed query execution and scheduling</li>\n<li>Vectorized execution engine</li>\n<li>Data security</li>\n<li>Resource management</li>\n<li>Transaction coordination</li>\n<li>Efficient storage structures (encodings, indexes)</li>\n<li>Automatic physical data optimization</li>\n</ul>\n<p>Your responsibilities will include driving requirements clarity and design decisions for ambiguous problems, producing technical design documents and project plans, developing new features, mentoring more junior engineers, testing and rolling out to production, and monitoring.</p>\n<p>We look for a passion for database systems, storage systems, distributed systems, language design, or performance optimization. You should be comfortable working towards a multi-year vision with incremental deliverables, be customer-oriented and focused on having an impact, and have 7+ years of experience working in a related system. A PhD in databases or distributed systems is optional.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8d587b18-3f1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8012818002","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimization","query compilation and optimization","distributed query execution and scheduling","vectorized execution engine","data security","resource management","transaction coordination","efficient storage structures","automatic physical data optimization"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:17.232Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Belgrade, Serbia"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimization, query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures, automatic physical data optimization"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_061e824c-343"},"title":"Software Engineer: Distributed Systems (Infrastructure)","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>Responsibilities</p>\n<p>As a Software Engineer: Distributed Systems, you will be part of a Resiliency Organization responsible for the core services that power Cloudflare&#39;s global operations. We are looking for engineers to join the Infrastructure Intelligence team and shape the transition toward model-driven network orchestration.</p>\n<p>The team is building a cutting-edge &#39;Maintenance Coordination System&#39;, powered by an infrastructure dependency graph of one of the world&#39;s largest physical networks. This is a foundational step towards designing intelligent, autonomous systems that will transform the orchestration of Cloudflare&#39;s network.</p>\n<p>It forms the basis of many future projects to build the core data structures and services required to ensure our network optimization, network forecasting, and capacity planning are the state of the art. By creating the robust primitives for global coordination today, you will be enabling the next generation of data-driven infrastructure at Cloudflare.</p>\n<p>This is a unique opportunity to work on complex, globally distributed systems which underpin all Cloudflare products.</p>\n<p>Technologies we use:</p>\n<ul>\n<li>Cloudflare Workers, Workers KV, R2, and Durable Objects</li>\n</ul>\n<ul>\n<li>Kubernetes</li>\n</ul>\n<ul>\n<li>Go, Typescript, Python</li>\n</ul>\n<ul>\n<li>For service monitoring we use Prometheus, Grafana and Sentry</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>A degree in Computer Science, Engineering, Mathematics, Statistics or related field; OR have relevant background/experience to the field.</li>\n</ul>\n<ul>\n<li>Programming experience in Go, or similar languages</li>\n</ul>\n<ul>\n<li>Experience in designing and implementing secure and highly-available distributed systems</li>\n</ul>\n<ul>\n<li>Experience (and love) for debugging to ensure the system works in all cases</li>\n</ul>\n<ul>\n<li>Experience with a continuous integration workflow and using source control (we use git)</li>\n</ul>\n<ul>\n<li>Experience with continuous delivery and deployment of a k8s hosted application</li>\n</ul>\n<ul>\n<li>Understanding of security issues and responsibilities</li>\n</ul>\n<ul>\n<li>Experience with monitoring, alerting and debugging high volume production systems</li>\n</ul>\n<ul>\n<li>Fluent in analyses of data sets such as logs</li>\n</ul>\n<ul>\n<li>Strong English language oral and written communications skills</li>\n</ul>\n<ul>\n<li>Designing and building APIs</li>\n</ul>\n<ul>\n<li>Experience with the Cloudflare development stack is a plus</li>\n</ul>\n<p>Examples of desirable skills, knowledge and experience</p>\n<ul>\n<li>At least 4 years of hands-on software development experience on meaningfully complex systems.</li>\n</ul>\n<ul>\n<li>Experience with graph theory and building services for graph generation, storage and retrieval.</li>\n</ul>\n<ul>\n<li>An understanding of the systems architecture required to scale machine learning model-driven decision engines in a production environment</li>\n</ul>\n<ul>\n<li>Experience building both backend systems and frontend widgets.</li>\n</ul>\n<ul>\n<li>Ability to contribute to planning, development, and execution to meet commitments and deliver with predictability.</li>\n</ul>\n<ul>\n<li>Experience implementing tools, processes, internal instrumentation, and methodologies.</li>\n</ul>\n<ul>\n<li>Comfortable working on projects with tight deadlines and short release cycles.</li>\n</ul>\n<ul>\n<li>Strong verbal and written English language skills.</li>\n</ul>\n<ul>\n<li>Experience with DCIM, CMDB, IPAM, and other Data Center and Asset Lifecycle Management tools is a plus.</li>\n</ul>\n<ul>\n<li>Experience with data ingestion and analysis - pulling metrics from hundreds of edge data centers.</li>\n</ul>\n<p>Compensation</p>\n<p>For Washington D.C. based hires: Estimated annual salary of $140,000 - 172,000.</p>\n<p>Equity</p>\n<p>This role is eligible to participate in Cloudflare&#39;s equity plan.</p>\n<p>Benefits</p>\n<p>Cloudflare offers a complete package of benefits and programs to support you and your family. Our benefits programs can help you pay health care expenses, support caregiving, build capital for the future and make life a little easier and fun!</p>\n<p>The below is a description of our benefits for employees in the United States, and benefits may vary for employees based outside the U.S.</p>\n<p>Health &amp; Welfare Benefits</p>\n<ul>\n<li>Medical/Rx Insurance</li>\n</ul>\n<ul>\n<li>Dental Insurance</li>\n</ul>\n<ul>\n<li>Vision Insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Accounts</li>\n</ul>\n<ul>\n<li>Commuter Spending Accounts</li>\n</ul>\n<ul>\n<li>Fertility &amp; Family Forming Benefits</li>\n</ul>\n<ul>\n<li>On-demand mental health support and Employee Assistance Program</li>\n</ul>\n<ul>\n<li>Global Travel Medical Insurance</li>\n</ul>\n<p>Financial Benefits</p>\n<ul>\n<li>Short and Long Term Disability Insurance</li>\n</ul>\n<ul>\n<li>Life &amp; Accident Insurance</li>\n</ul>\n<ul>\n<li>401(k) Retirement Savings Plan</li>\n</ul>\n<ul>\n<li>Employee Stock Participation Plan</li>\n</ul>\n<p>Time Off</p>\n<ul>\n<li>Flexible paid time off covering vacation and sick leave</li>\n</ul>\n<ul>\n<li>Leave programs, including parental, pregnancy health, medical, and bereavement leave</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We&#39;re not just a highly ambitious, large-scale technology company. We&#39;re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare&#39;s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever built.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_061e824c-343","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7088208","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Programming experience in Go, or similar languages","Experience in designing and implementing secure and highly-available distributed systems","Experience (and love) for debugging to ensure the system works in all cases","Experience with a continuous integration workflow and using source control (we use git)","Experience with continuous delivery and deployment of a k8s hosted application","Understanding of security issues and responsibilities","Experience with monitoring, alerting and debugging high volume production systems","Fluent in analyses of data sets such as logs","Strong English language oral and written communications skills","Designing and building APIs","Experience with the Cloudflare development stack is a plus"],"x-skills-preferred":["At least 4 years of hands-on software development experience on meaningfully complex systems","Experience with graph theory and building services for graph generation, storage and retrieval","An understanding of the systems architecture required to scale machine learning model-driven decision engines in a production environment","Experience building both backend systems and frontend widgets","Ability to contribute to planning, development, and execution to meet commitments and deliver with predictability","Experience implementing tools, processes, internal instrumentation, and methodologies","Comfortable working on projects with tight deadlines and short release cycles","Strong verbal and written English language skills","Experience with DCIM, CMDB, IPAM, and other Data Center and Asset Lifecycle Management tools is a plus","Experience with data ingestion and analysis - pulling metrics from hundreds of edge data centers"],"datePosted":"2026-04-18T15:45:07.400Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Programming experience in Go, or similar languages, Experience in designing and implementing secure and highly-available distributed systems, Experience (and love) for debugging to ensure the system works in all cases, Experience with a continuous integration workflow and using source control (we use git), Experience with continuous delivery and deployment of a k8s hosted application, Understanding of security issues and responsibilities, Experience with monitoring, alerting and debugging high volume production systems, Fluent in analyses of data sets such as logs, Strong English language oral and written communications skills, Designing and building APIs, Experience with the Cloudflare development stack is a plus, At least 4 years of hands-on software development experience on meaningfully complex systems, Experience with graph theory and building services for graph generation, storage and retrieval, An understanding of the systems architecture required to scale machine learning model-driven decision engines in a production environment, Experience building both backend systems and frontend widgets, Ability to contribute to planning, development, and execution to meet commitments and deliver with predictability, Experience implementing tools, processes, internal instrumentation, and methodologies, Comfortable working on projects with tight deadlines and short release cycles, Strong verbal and written English language skills, Experience with DCIM, CMDB, IPAM, and other Data Center and Asset Lifecycle Management tools is a plus, Experience with data ingestion and analysis - pulling metrics from hundreds of edge data centers"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_65bcd8f4-7e6"},"title":"Staff Software Engineer - Database Engine Internals","description":"<p>Our mission at Databricks is to radically simplify the whole data lifecycle from ingestion to ETL, BI, and all the way up to ML/AI with a unified platform.</p>\n<p>To achieve this goal, we believe the data warehouse architecture as we know it today will be replaced by a new architectural pattern, Lakehouse (CIDR 2021 paper), open platforms that unify data warehousing and advanced analytics.</p>\n<p>A critical part of realizing this vision is the next generation (decoupled) query engine and structured storage system that can outperform specialised data warehouses in relational query performance, yet retain the expressiveness and of general purpose systems such as Apache Spark™ to support diverse workloads ranging from ETL to data science.</p>\n<p>As part of this team, you will be working in one or more of the following areas to design and implement these next gen systems that leapfrog state-of-the-art:</p>\n<ul>\n<li>Query compilation and optimisation</li>\n<li>Distributed query execution and scheduling</li>\n<li>Vectorised execution engine</li>\n<li>Data security</li>\n<li>Resource management</li>\n<li>Transaction coordination</li>\n<li>Efficient storage structures (encodings, indexes)</li>\n<li>Automatic physical data optimisation</li>\n</ul>\n<p>We look for:</p>\n<ul>\n<li>A passion for database systems, storage systems, distributed systems, language design, or performance optimisation</li>\n<li>Experience working towards a multi-year vision with incremental deliverables</li>\n<li>Motivated by delivering customer value and impact</li>\n<li>8+ years of experience working in a related system (preferred)</li>\n</ul>\n<p>Pay Range Transparency</p>\n<p>Databricks is committed to fair and equitable compensation practices. The pay range(s) for this role is listed below and represents the expected salary range for non-commissionable roles or on-target earnings for commissionable roles.</p>\n<p>Actual compensation packages are based on several factors that are unique to each candidate, including but not limited to job-related skills, depth of experience, relevant certifications and training, and specific work location.</p>\n<p>Based on the factors above, Databricks anticipates utilising the full width of the range.</p>\n<p>The total compensation package for this position may also include eligibility for annual performance bonus, equity, and the benefits listed above.</p>\n<p>For more information regarding which range your location is in visit our page here.</p>\n<p>Local Pay Range $192,000-$260,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_65bcd8f4-7e6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/5646866002","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$192,000-$260,000 USD","x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimisation","query compilation","optimisation","distributed query execution","scheduling","vectorised execution engine","data security","resource management","transaction coordination","efficient storage structures","encodings","indexes"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:40.153Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimisation, query compilation, optimisation, distributed query execution, scheduling, vectorised execution engine, data security, resource management, transaction coordination, efficient storage structures, encodings, indexes","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":192000,"maxValue":260000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_369bdc50-553"},"title":"Member of Technical Staff - Real-Time Storage","description":"<p>We&#39;re looking for exceptional storage &amp; database engineers to join us on a new project to deeply integrate xAI&#39;s advanced AI infrastructure into a platform used by around 600 million users every month.</p>\n<p>Our team is building xAI&#39;s new storage tier that powers training, inference, recommendations, and real-time data extraction. This includes an exabyte-scale S3-compatible object store, a multi-tenant transactional key/value store, a caching tier, a scalable vector database, and a few other specialised databases.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and launch to production new features and improvements aimed at unifying common components across the storage systems</li>\n<li>Dive into performance issues and work with customers and deliver solutions to cater to customers&#39; latency, availability, and data durability requirements</li>\n<li>Lead and drive incident responses and recovery with your peers. Review and contribute to incident postmortems and hold a high bar for the same</li>\n<li>Work in a collaborative environment and uplevel your peers by doing mentoring, code and design reviews</li>\n<li>Be open to develop new skills and learn on the job as we navigate new technology spaces</li>\n<li>Ground your decisions with data and reasoning and adapt to new information to make informed choices</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>Have at least 7 years of experience in software development, plus 2+ years building storage systems or databases</li>\n<li>Obsessed with reliability, performance, and quality</li>\n<li>Proficient in high-performance C++, Rust, or JVM-based languages</li>\n<li>Expertise in building, running, and operating scalable and resilient distributed systems</li>\n</ul>\n<p>Compensation and Benefits:</p>\n<p>$180,000 - $440,000 USD</p>\n<p>Base salary is just one part of our total rewards package at xAI, which also includes equity, comprehensive medical, vision, and dental coverage, access to a 401(k) retirement plan, short &amp; long-term disability insurance, life insurance, and various other discounts and perks.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_369bdc50-553","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.xai.com/","logo":"https://logos.yubhub.co/xai.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/4805886007","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$180,000 - $440,000 USD","x-skills-required":["C++","Rust","JVM-based languages","Distributed systems","Storage systems","Database systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:35.225Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Palo Alto, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"C++, Rust, JVM-based languages, Distributed systems, Storage systems, Database systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":440000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b00b5e68-8bf"},"title":"Principal Software Developer I / II - Storage Engine - Elasticsearch","description":"<p>We&#39;re looking for a Principal Software Developer I or II to join the Elasticsearch - Storage Engine team. This globally-distributed, completely remote team of senior engineers is responsible for delivering the latest innovations in logs and metrics management.</p>\n<p>This role includes providing technical vision and direction for building solutions that provide optimized storage and efficient data querying and indexing. This role requires related past technical experience in addition to the ability to work cross-organisation.</p>\n<p>Our company is distributed by intention. We hire the best developers we can find wherever they are, whoever they are. We collaborate across continents every day over email, GitHub, Zoom, and Slack. At our best, we write fast, scalable and intuitive software. We believe that the best way to do that is to empower individual engineers, code review every change, decide big things by consensus, and strive for incremental improvements.</p>\n<p>As a Principal Software Developer, you will lead cross-organisational initiatives to produce an industry-leading Timeseries solution offering. You will contribute to Elasticsearch full time, building data-intensive new features and fixing intriguing bugs, all while making the code easier to understand. Sometimes you&#39;ll need to implement a data structure or algorithm in the code base. And there will be times when you&#39;ll need to get close to the operating system and hardware.</p>\n<p>You will work with a globally distributed team of experienced engineers focused on the logs and metrics capabilities of Elasticsearch. You will be an expert in several areas of Elasticsearch and everyone will turn to you when they have a question about them. You&#39;ll improve those areas based on your questions and your instincts.</p>\n<p>You will work with community members from all over the world on issues and pull requests, sometimes triaging them and handing them off to other experts and sometimes handling them yourself. You will write idiomatic modern Java -- Elasticsearch is 99.8% Java!</p>\n<p>We&#39;re looking for someone who has implemented novel techniques to efficiently index, store and query Timeseries data. You should have strong technical leadership skills, work with a high level of autonomy, and be able to take on projects and guide them from beginning to end. This covers both technical design and working with other engineers to develop needed components.</p>\n<p>You should have strong skills in core Java and be conversant in the standard library of data structures and concurrency constructs, as well as newer features like lambdas. You should have a strong desire to optimise and make use of the most efficient data structures and algorithms.</p>\n<p>You&#39;re comfortable developing collaboratively. Giving and receiving feedback on code and approaches and APIs is hard! Bonus points if you&#39;ve collaborated over the internet because that&#39;s harder. Double bonus points for asynchronous collaboration over the internet. That&#39;s even harder but we do it anyway because it&#39;s the best way we know how to build software.</p>\n<p>You&#39;ve used several data storage technologies like Elasticsearch, Solr, PostgreSQL, MongoDB, or Cassandra and have some idea how they work and why they work that way. You have excellent verbal and written communication skills. Like we said, collaborating on the internet is hard. We try to be respectful, empathetic, and trusting in all of our interactions. And we&#39;d expect that from you too.</p>\n<p>Bonus points if you&#39;ve built things with Elasticsearch before. Bonus points if you&#39;ve worked with open source projects and are familiar with different styles of source control workflow and continuous integration.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b00b5e68-8bf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Elastic","sameAs":"https://www.elastic.co/","logo":"https://logos.yubhub.co/elastic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/elastic/jobs/7348825","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$192,500-$304,500 CAD","x-skills-required":["Java","Elasticsearch","Timeseries data","Data structures","Concurrency constructs","Lambdas","Data storage technologies"],"x-skills-preferred":["Open source projects","Source control workflow","Continuous integration"],"datePosted":"2026-04-18T15:44:28.872Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Elasticsearch, Timeseries data, Data structures, Concurrency constructs, Lambdas, Data storage technologies, Open source projects, Source control workflow, Continuous integration","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":192500,"maxValue":304500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_982dd81e-416"},"title":"Principal Database Engineer, Data Engineering","description":"<p>As a Principal Database Engineer, you&#39;ll design and lead the evolution of the PostgreSQL backbone that powers GitLab.com and thousands of self-managed enterprise deployments. You&#39;ll solve critical challenges around uncontrolled data growth, complex upgrades and migrations, and always-on reliability at global scale, creating the database patterns and platforms that keep GitLab fast, resilient, and cost efficient as usage grows.</p>\n<p>You&#39;ll architect scalable, distributed database solutions, build proactive health and reliability frameworks, and drive adoption of modern database technologies and data stores that improve both product capabilities and production stability. Working hands-on in the codebase and partnering closely with product and infrastructure teams, you&#39;ll turn long-term database strategy into incremental, customer-visible improvements, shift incident response from reactive to proactive, and help define GitLab&#39;s next-generation data architecture, including sharding and multi-database support.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Lead the architecture and strategy for GitLab.com&#39;s PostgreSQL infrastructure, designing scalable, resilient solutions for both SaaS and self-managed deployments.</li>\n</ul>\n<ul>\n<li>Build proactive database health and reliability frameworks using continuous monitoring, automated remediation, and predictive analytics to prevent customer-impacting incidents.</li>\n</ul>\n<ul>\n<li>Drive database best practices across engineering by guiding schema design, migrations, and query optimization, and by creating self-service tools and guardrails for product teams.</li>\n</ul>\n<ul>\n<li>Own end-to-end observability for database systems, designing symptom-based monitoring, leading incident response, and turning learnings into automated, repeatable workflows.</li>\n</ul>\n<ul>\n<li>Shape the evolution of GitLab’s database platform by evaluating and implementing modern database technologies and data stores that improve reliability, performance, and product capabilities.</li>\n</ul>\n<ul>\n<li>Design solutions and patterns that address uncontrolled data growth, cost efficiency, sharding, multi-database support, and other next-generation data architecture needs.</li>\n</ul>\n<ul>\n<li>Collaborate closely with product and infrastructure teams to align product decisions with platform constraints and priorities, breaking down long-term goals into incremental, customer-visible outcomes.</li>\n</ul>\n<ul>\n<li>Contribute directly to the codebase to prototype and ship working solutions, maintain technical credibility, and deep-dive into complex production issues when needed.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Experience architecting, operating, and optimizing PostgreSQL in large-scale, distributed production environments with high availability and disaster recovery requirements.</li>\n</ul>\n<ul>\n<li>Deep knowledge of PostgreSQL internals, including the query planner, write-ahead logging, vacuum processes, and storage engine behavior.</li>\n</ul>\n<ul>\n<li>Background designing and maintaining highly distributed database platforms with automated failover, robust monitoring, and self-healing capabilities.</li>\n</ul>\n<ul>\n<li>Hands-on coding skills and comfort working across the stack, from low-level database and search systems to backend and frontend services.</li>\n</ul>\n<ul>\n<li>Familiarity with infrastructure-as-code, GitOps practices, security hardening, and site reliability engineering principles applied to database operations.</li>\n</ul>\n<ul>\n<li>Ability to debug complex, cross-system issues, translate findings into durable technical solutions, and turn incident learnings into repeatable automation.</li>\n</ul>\n<ul>\n<li>Experience influencing technical direction across multiple teams, providing practical guidance on migrations, query optimization, and database best practices.</li>\n</ul>\n<ul>\n<li>Openness to collaborating with people from diverse technical backgrounds, with a focus on clear communication, shared ownership, and learning transferable skills.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_982dd81e-416","directApply":true,"hiringOrganization":{"@type":"Organization","name":"GitLab","sameAs":"https://about.gitlab.com/","logo":"https://logos.yubhub.co/about.gitlab.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gitlab/jobs/8231379002","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$157,900-$338,400 USD","x-skills-required":["PostgreSQL","database architecture","data engineering","infrastructure-as-code","GitOps","security hardening","site reliability engineering","database operations","query optimization","schema design","migrations","query planning","write-ahead logging","vacuum processes","storage engine behavior"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:15.402Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, EMEA; Remote, North America"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"PostgreSQL, database architecture, data engineering, infrastructure-as-code, GitOps, security hardening, site reliability engineering, database operations, query optimization, schema design, migrations, query planning, write-ahead logging, vacuum processes, storage engine behavior","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":157900,"maxValue":338400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_696d40f3-265"},"title":"Senior/Principal Product Manager - Durable Objects","description":"<p>About Us</p>\n<p>At Cloudflare, we&#39;re on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>As a Senior/Principal Product Manager - Durable Objects, you&#39;ll be responsible for leading Durable Objects, a powerful compute &amp; storage primitive used by customers and many of Cloudflare&#39;s own services as part of Cloudflare&#39;s Developer Platform. Durable Objects unlock stateful Workers applications on Cloudflare&#39;s global network. Many of Cloudflare&#39;s products, Agents SDK, Containers, Workflows, D1, etc, are built on top of Durable Objects.</p>\n<p>This is a role with high agency and accountability: you&#39;ll be the sole PM for the products you&#39;re responsible for, and you&#39;ll work closely with your peers across the team to ensure we&#39;re shipping great developer experiences and not silos.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Drive the strategy, roadmap &amp; (most importantly) ship great products that help developers build reliable, high performance full-stack applications on Cloudflare.</li>\n</ul>\n<ul>\n<li>Lead projects from ideation through launch, and continuing through ongoing maintenance and support</li>\n</ul>\n<ul>\n<li>Develop and nurture relationships with engineering and coordinate closely to ensure successful delivery of product features</li>\n</ul>\n<ul>\n<li>Talk to developers (probably the most important)! You&#39;ll need to be crafty and develop a playbook to find the right users to talk to. This might include diving into our BI tools to find specific cohorts that are worth talking to, engaging on our developer discord, or heading straight to Twitter or Hacker News to engage and build relationships with the community.</li>\n</ul>\n<ul>\n<li>Partner with other departments such as Product Marketing, Marketing, Community &amp; Developer Relations, Sales, and Customer Support to drive adoption with and gather feedback from customers and prospects</li>\n</ul>\n<ul>\n<li>Ruthlessly prioritize across a complex environment of diverse customers, stakeholders and objectives</li>\n</ul>\n<ul>\n<li>Represent the company externally as needed, including meetings with customers and prospects and analysis</li>\n</ul>\n<ul>\n<li>Delve into market research to drive new investment opportunities in the developer space</li>\n</ul>\n<p>Examples Of Desirable Skills, Knowledge And Experience</p>\n<ul>\n<li>Outstanding communication and collaboration skills and are able to work with a diverse group, get alignment, and drive the product forward.</li>\n</ul>\n<ul>\n<li>Execution focused and emphasize getting things done while paying attention to important details.</li>\n</ul>\n<ul>\n<li>A desire to sweat the details and ship great end-to-end developer experiences. APIs, UIs, docs: all of these are part of the product.</li>\n</ul>\n<ul>\n<li>Proven track record of shipping developer-facing products from ideation to launch. You should know what it means to be a developer and what gets in the way of building production services.</li>\n</ul>\n<ul>\n<li>A love for talking to developers at organizations both large and small on a daily basis so that you can build and ship products that solve real problems</li>\n</ul>\n<ul>\n<li>Demonstrated experience with transactional databases or storage products (NoSQL or SQL, building on or building them directly), and a desire to become a domain expert in the space as quickly as humanly possible.</li>\n</ul>\n<ul>\n<li>A focus on the developer: you&#39;ll be working directly with customers, and often be a part of the customer&#39;s team in helping them get their initial implementation off the ground</li>\n</ul>\n<ul>\n<li>BS/MS in a technology field</li>\n</ul>\n<ul>\n<li>Professional product management experience (5-8 years), previous founder experience or demonstrated experience on an engineering team shipping user-facing products.</li>\n</ul>\n<ul>\n<li>Strong technical abilities. You should be intimately familiar with modern software development practices used to build and deploy applications. You are capable of wading into the details &amp; understanding engineering decisions&#39; impact on the product, but can also step back and take a 30k foot view.</li>\n</ul>\n<ul>\n<li>Experience using developer tools: you should have a favorite text editor, be familiar with composing CLI tools, and be the first user of the products you’re shipping.</li>\n</ul>\n<ul>\n<li>Programming experience (JavaScript/TypeScript, Rust, Go, etc.) and/or experience with SQL dialects.</li>\n</ul>\n<p>Benefits</p>\n<ul>\n<li>Compensation may be adjusted depending on work location.</li>\n</ul>\n<ul>\n<li>For Bay Area based hires: Estimated annual salary of $200k to $240k Equity</li>\n</ul>\n<ul>\n<li>This role is eligible to participate in Cloudflare’s equity plan.</li>\n</ul>\n<ul>\n<li>Cloudflare offers a complete package of benefits and programs to support you and your family.</li>\n</ul>\n<ul>\n<li>Our benefits programs can help you pay health care expenses, support caregiving, build capital for the future and make life a little easier and fun!</li>\n</ul>\n<ul>\n<li>The below is a description of our benefits for employees in the United States, and benefits may vary for employees based outside the U.S.</li>\n</ul>\n<ul>\n<li>Health &amp; Welfare Benefits</li>\n</ul>\n<ul>\n<li>Medical/Rx Insurance</li>\n</ul>\n<ul>\n<li>Dental Insurance</li>\n</ul>\n<ul>\n<li>Vision Insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Accounts</li>\n</ul>\n<ul>\n<li>Commuter Spending Accounts</li>\n</ul>\n<ul>\n<li>Fertility &amp; Family Forming Benefits</li>\n</ul>\n<ul>\n<li>On-demand mental health support and Employee Assistance Program</li>\n</ul>\n<ul>\n<li>Global Travel Medical Insurance</li>\n</ul>\n<ul>\n<li>Financial Benefits</li>\n</ul>\n<ul>\n<li>Short and Long Term Disability Insurance</li>\n</ul>\n<ul>\n<li>Life &amp; Accident Insurance</li>\n</ul>\n<ul>\n<li>401(k) Retirement Savings Plan</li>\n</ul>\n<ul>\n<li>Employee Stock Participation Plan</li>\n</ul>\n<ul>\n<li>Time Off</li>\n</ul>\n<ul>\n<li>Flexible paid time off covering vacation and sick leave</li>\n</ul>\n<ul>\n<li>Leave programs, including parental, pregnancy health, medical, and bereavement leave</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<ul>\n<li>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul.</li>\n</ul>\n<ul>\n<li>Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</li>\n</ul>\n<ul>\n<li>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</li>\n</ul>\n<ul>\n<li>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration.</li>\n</ul>\n<ul>\n<li>Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</li>\n</ul>\n<ul>\n<li>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</li>\n</ul>\n<ul>\n<li>Here</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_696d40f3-265","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7731886","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["communication and collaboration skills","execution focused","desire to sweat the details","proven track record of shipping developer-facing products","demonstrated experience with transactional databases or storage products","focus on the developer","BS/MS in a technology field","professional product management experience","strong technical abilities","experience using developer tools","programming experience","experience with SQL dialects"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:09.698Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"communication and collaboration skills, execution focused, desire to sweat the details, proven track record of shipping developer-facing products, demonstrated experience with transactional databases or storage products, focus on the developer, BS/MS in a technology field, professional product management experience, strong technical abilities, experience using developer tools, programming experience, experience with SQL dialects"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d8e9f3a9-77b"},"title":"Intermediate Backend Engineer (C), Tenant Scale: Git","description":"<p>As a Backend Engineer on the Tenant Scale: Git team, you&#39;ll contribute to improving one of the most widely used foundations of modern software development: Git. Git is at the core of how developers collaborate, and this role focuses on making Git and Gitaly more capable, reliable, and efficient for GitLab and the people who use our platform.</p>\n<p>You&#39;ll participate in architectural discussions and technical decisions related to Git and Gitaly, helping drive implementation choices that improve correctness, performance, and maintainability. You&#39;ll contribute features, bug fixes, and performance improvements to upstream Git in line with team and community goals, delivering changes that improve repository access and reliability for users.</p>\n<p>Adapt Gitaly to make effective use of Git capabilities, including integrating newly available features to improve scalability, efficiency, and long-term maintainability. Connect discussions in the open source Git project with GitLab&#39;s product direction and engineering work, helping align upstream contributions with product and platform needs.</p>\n<p>Scope tasks, estimate effort, and describe implementation plans that support the team&#39;s priorities and enable predictable delivery of technical work. Test and validate the features you build and integrate, with a focus on correctness and reliability to reduce regressions and support stable production use.</p>\n<p>Collaborate with team members, contributors, and the Git ecosystem. Represent GitLab as a constructive participant in the open source ecosystem, building productive relationships that support ongoing collaboration with the Git community.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d8e9f3a9-77b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"GitLab","sameAs":"https://about.gitlab.com/","logo":"https://logos.yubhub.co/about.gitlab.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gitlab/jobs/8497793002","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Git internals","C programming language","Go for backend development","Linux internals","Open source projects","Distributed systems","Storage formats","Graph theory","Highly available production environments"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:43:30.153Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, India"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Git internals, C programming language, Go for backend development, Linux internals, Open source projects, Distributed systems, Storage formats, Graph theory, Highly available production environments"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_90b4c10b-948"},"title":"Staff Software Engineer - Database Engine Internals","description":"<p>Our mission at Databricks is to simplify the data lifecycle from ingestion to ETL, BI, and all the way up to ML/AI with a unified platform.</p>\n<p>To achieve this goal, we believe the data warehouse architecture as we know it today will be replaced by a new architectural pattern, Lakehouse, open platforms that unify data warehousing and advanced analytics.</p>\n<p>A critical part of realizing this vision is the next generation (decoupled) query engine and structured storage system that can outperform specialized data warehouses in relational query performance, yet retain the expressiveness and of general purpose systems such as Apache Spark to support diverse workloads ranging from ETL to data science.</p>\n<p>As part of this team, you will be working in one or more of the following areas to design and implement these next gen systems that leapfrog state-of-the-art:</p>\n<ul>\n<li>Query compilation and optimization</li>\n<li>Distributed query execution and scheduling</li>\n<li>Vectorized execution engine</li>\n<li>Data security</li>\n<li>Resource management</li>\n<li>Transaction coordination</li>\n<li>Efficient storage structures (encodings, indexes)</li>\n<li>Automatic physical data optimization</li>\n</ul>\n<p>We look for individuals with a passion for database systems, storage systems, distributed systems, language design, or performance optimization. You should have experience working towards a multi-year vision with incremental deliverables, motivated by delivering customer value and impact.</p>\n<p>The pay range for this role is $192,000-$260,000 USD, and the total compensation package may also include eligibility for annual performance bonus, equity, and benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_90b4c10b-948","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/6544386002","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$192,000-$260,000 USD","x-skills-required":["database systems","storage systems","distributed systems","language design","performance optimization","query compilation and optimization","distributed query execution and scheduling","vectorized execution engine","data security","resource management","transaction coordination","efficient storage structures"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:43:13.600Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database systems, storage systems, distributed systems, language design, performance optimization, query compilation and optimization, distributed query execution and scheduling, vectorized execution engine, data security, resource management, transaction coordination, efficient storage structures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":192000,"maxValue":260000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2fe8215c-605"},"title":"Senior Software Engineer, Storage Infrastructure","description":"<p>About Us</p>\n<p>At Cloudflare, we&#39;re on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>Cloudflare protects and accelerates any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>Emerging Technologies &amp; Incubation (ETI)</p>\n<p>ETI is where new and bold products are built and released within Cloudflare. Rather than being constrained by the structures which make Cloudflare a massively successful business, we are able to leverage them to deliver entirely new tools and products to our customers. Cloudflare&#39;s edge and network make it possible to solve problems at massive scale and efficiency which would be impossible for almost any other organization.</p>\n<p>About the Team</p>\n<p>ETI&#39;s Storage Infrastructure team is responsible for the core storage layer that underpins many of ETI&#39;s stateful services. Our scope ranges from managing the physical hardware to operating the distributed databases and storage systems built upon it. We run this infrastructure globally across Cloudflare&#39;s network, which presents unique and complex engineering puzzles. We navigate efficiently expanding storage capacity, optimizing rebuild operations, and coordinating operations across failure domains to uphold durability.</p>\n<p>While other service teams focus on product development, our mission is to ensure the underlying storage is reliable, performant, and scalable. You&#39;ll be joining a highly motivated team that is building the next generation of distributed storage services.</p>\n<p>Responsibilities</p>\n<p>In this role, you will help build and operate the next generation of globally distributed storage systems. You will own your code from inception to release, delivering solutions at all layers of the stack. On any given day, you might write a design document for a new provisioning system, model failure domain dependencies across edge locations, benchmark new storage hardware, build standardized observability and runbooks for distributed database clusters, or automate operational toil through purpose-built tooling and intelligent automation.</p>\n<p>You can expect to interact with a variety of languages and technologies including Rust, Go, Saltstack, and Terraform.</p>\n<p>Examples of desirable skills, knowledge, and experience</p>\n<ul>\n<li>Strong programming skills with languages like Rust, Go, or Python</li>\n<li>A solid understanding of distributed systems concepts such as consistency, consensus, data replication, fault tolerance, and partition tolerance</li>\n<li>Experience with distributed databases and storage systems</li>\n<li>Experience with infrastructure configuration tooling and infrastructure as code</li>\n<li>Familiarity with storage fundamentals: block devices, filesystems, SSD characteristics</li>\n<li>Experience building and maintaining high-throughput, low-latency systems</li>\n<li>Understanding of network fundamentals as they relate to distributed storage -- bandwidth constraints, latency tradeoffs, cross-datacenter replication</li>\n<li>Strong written and verbal communication skills and ability to explain technical decisions clearly</li>\n<li>Comfortable operating in fast-paced environments with tight deadlines and evolving priorities</li>\n</ul>\n<p>Benefits</p>\n<p>Cloudflare offers a complete package of benefits and programs to support you and your family. Our benefits programs can help you pay health care expenses, support caregiving, build capital for the future and make life a little easier and fun!</p>\n<p>The below is a description of our benefits for employees in the United States, and benefits may vary for employees based outside the U.S.</p>\n<p>Health &amp; Welfare Benefits</p>\n<ul>\n<li>Medical/Rx Insurance</li>\n<li>Dental Insurance</li>\n<li>Vision Insurance</li>\n<li>Flexible Spending Accounts</li>\n<li>Commuter Spending Accounts</li>\n<li>Fertility &amp; Family Forming Benefits</li>\n<li>On-demand mental health support and Employee Assistance Program</li>\n<li>Global Travel Medical Insurance</li>\n</ul>\n<p>Financial Benefits</p>\n<ul>\n<li>Short and Long Term Disability Insurance</li>\n<li>Life &amp; Accident Insurance</li>\n<li>401(k) Retirement Savings Plan</li>\n<li>Employee Stock Participation Plan</li>\n</ul>\n<p>Time Off</p>\n<ul>\n<li>Flexible paid time off covering vacation and sick leave</li>\n<li>Leave programs, including parental, pregnancy health, medical, and bereavement leave</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We&#39;re not just a highly ambitious, large-scale technology company. We&#39;re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo</p>\n<p>Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare&#39;s enterprise customers--at no cost.</p>\n<p>Athenian Project</p>\n<p>In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1</p>\n<p>We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released. Here&#39;s the deal - we don&#39;t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you&#39;d like to be a part of? We&#39;d love to hear from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2fe8215c-605","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7629805","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Rust","Go","Python","Distributed systems","Consistency","Consensus","Data replication","Fault tolerance","Partition tolerance","Distributed databases","Storage systems","Infrastructure configuration tooling","Infrastructure as code","Storage fundamentals","Block devices","Filesystems","SSD characteristics","High-throughput systems","Low-latency systems","Network fundamentals","Bandwidth constraints","Latency tradeoffs","Cross-datacenter replication"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:42:33.190Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Rust, Go, Python, Distributed systems, Consistency, Consensus, Data replication, Fault tolerance, Partition tolerance, Distributed databases, Storage systems, Infrastructure configuration tooling, Infrastructure as code, Storage fundamentals, Block devices, Filesystems, SSD characteristics, High-throughput systems, Low-latency systems, Network fundamentals, Bandwidth constraints, Latency tradeoffs, Cross-datacenter replication"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_15bec9eb-375"},"title":"Staff Software Engineer - Backend","description":"<p>We are seeking a Staff Software Engineer - Backend to join our London site and contribute to our multi-year journey to build the best Lakehouse Platform. As a founding member of this team, you will be involved in the entire development cycle and exemplify all core Databricks values. Your impact will be significant, as you will work on challenges such as distributed systems, at-scale service architecture and monitoring, workflow orchestration, and developer experience. You will also build reliable, secure and high performance services and client libraries for storing and accessing humongous amounts of data on cloud storage backends.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Developing product features that empower our customers to easily store and access their data</li>\n<li>Solving reliability problems related to Lakebase</li>\n<li>Actively finding causes of downtime and systematically improving or removing root causes</li>\n<li>Helping the org define SLIs, meet SLOs, and drive long-term reliability improvements</li>\n</ul>\n<p>To succeed in this role, you will need:</p>\n<ul>\n<li>A BS degree (or higher) in Computer Science, or a related field</li>\n<li>8+ years of production level experience in one of: Java, Scala, C++, or similar language</li>\n<li>Experience developing large-scale distributed systems</li>\n<li>Experience working on a SaaS platform or with Service-Oriented Architectures</li>\n<li>Knowledge of SQL</li>\n</ul>\n<p>At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please click here.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_15bec9eb-375","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8374611002","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Scala","C++","distributed systems","at-scale service architecture and monitoring","workflow orchestration","developer experience","cloud storage backends","SQL"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:42:28.413Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Scala, C++, distributed systems, at-scale service architecture and monitoring, workflow orchestration, developer experience, cloud storage backends, SQL"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_18ae1499-b22"},"title":"Research Engineer, Discovery","description":"<p>As a Research Engineer on our team, you will work end-to-end across the whole model stack, identifying and addressing key infra blockers on the path to scientific AGI. Strong candidates should have familiarity with elements of language model training, evaluation, and inference and eagerness to quickly dive and get up to speed in areas they are not yet an expert on.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and implement large-scale infrastructure systems to support AI scientist training, evaluation, and deployment across distributed environments</li>\n<li>Identify and resolve infrastructure bottlenecks impeding progress toward scientific capabilities</li>\n<li>Develop robust and reliable evaluation frameworks for measuring progress towards scientific AGI</li>\n<li>Build scalable and performant VM/sandboxing/container architectures to safely execute long-horizon AI tasks and scientific workflows</li>\n<li>Collaborate to translate experimental requirements into production-ready infrastructure</li>\n<li>Develop large scale data pipelines to handle advanced language model training requirements</li>\n<li>Optimize large scale training and inference pipelines for stable and efficient reinforcement learning</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Have 6+ years of highly-relevant experience in infrastructure engineering with demonstrated expertise in large-scale distributed systems</li>\n<li>Are a strong communicator and enjoy working collaboratively</li>\n<li>Possess deep knowledge of performance optimization techniques and system architectures for high-throughput ML workloads</li>\n<li>Have experience with containerization technologies (Docker, Kubernetes) and orchestration at scale</li>\n<li>Have proven track record of building large-scale data pipelines and distributed storage systems</li>\n<li>Excel at diagnosing and resolving complex infrastructure challenges in production environments</li>\n<li>Can work effectively across the full ML stack from data pipelines to performance optimization</li>\n<li>Have experience collaborating with other researchers to scale experimental ideas</li>\n<li>Thrive in fast-paced environments and can rapidly iterate from experimentation to production</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Experience with language model training infrastructure and distributed ML frameworks (PyTorch, JAX, etc.)</li>\n<li>Background in building infrastructure for AI research labs or large-scale ML organizations</li>\n<li>Knowledge of GPU/TPU architectures and language model inference optimization</li>\n<li>Experience with cloud platforms (AWS, GCP) at enterprise scale</li>\n<li>Familiarity with VM and container orchestration</li>\n<li>Experience with workflow orchestration tools and experiment management systems</li>\n<li>History working with large scale reinforcement learning</li>\n<li>Comfort with large scale data pipelines (Beam, Spark, Dask, …)</li>\n</ul>\n<p>The annual compensation range for this role is $350,000-$850,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_18ae1499-b22","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4669581008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$350,000-$850,000 USD","x-skills-required":["large-scale distributed systems","containerization technologies (Docker, Kubernetes)","performance optimization techniques","system architectures for high-throughput ML workloads","data pipelines","distributed storage systems","ML frameworks (PyTorch, JAX, etc.)","GPU/TPU architectures","cloud platforms (AWS, GCP)","VM and container orchestration","workflow orchestration tools","experiment management systems","reinforcement learning","large scale data pipelines (Beam, Spark, Dask, …)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:41:42.408Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"large-scale distributed systems, containerization technologies (Docker, Kubernetes), performance optimization techniques, system architectures for high-throughput ML workloads, data pipelines, distributed storage systems, ML frameworks (PyTorch, JAX, etc.), GPU/TPU architectures, cloud platforms (AWS, GCP), VM and container orchestration, workflow orchestration tools, experiment management systems, reinforcement learning, large scale data pipelines (Beam, Spark, Dask, …)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":350000,"maxValue":850000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e994a45c-88a"},"title":"IT Systems Engineer","description":"<p>The IT Systems Team at xAI manages a dynamic and evolving virtual infrastructure providing secure, high-performance, and reliable corporate resources for business-critical users and applications.</p>\n<p>This role focuses on building, deploying, and managing internal Windows and Linux infrastructure while collaborating with other teams to develop scalable solutions.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Building, managing, and supporting complex application stacks on Windows and Linux Servers.</li>\n<li>Providing Windows and Linux troubleshooting and technical support.</li>\n<li>Deploy, configure and support VMware technologies, including vCenter, ESX and Horizon.</li>\n<li>Use Puppet Enterprise to automate and manage configurations.</li>\n<li>Working with other engineering teams and third-party vendors to implement scalable solutions.</li>\n<li>Ensuring compliance by collaborating with internal audit and compliance teams.</li>\n<li>Developing and maintaining standard operating procedures.</li>\n<li>Participating in a 24/7 on-call support rotation.</li>\n<li>Managing and monitoring all installed Windows and Linux systems and infrastructure.</li>\n<li>Installing, configuring, testing, and maintaining operating systems and system management tools.</li>\n<li>Proactively ensuring high levels of system availability and reliability.</li>\n<li>Monitoring and optimizing application performance, identifying bottlenecks, and collaborating with developers to implement solutions.</li>\n<li>Writing and maintaining custom scripts to improve system efficiency and automation.</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li>Deep understanding of infrastructure management and automation</li>\n<li>3 to 5 years of experience in the following:</li>\n<li>Systems Engineering for Corporate Infrastructure.</li>\n<li>Managing large-scale Windows and Linux server environments.</li>\n<li>Virtualization at scale.</li>\n<li>Configuration Management.</li>\n<li>Managing scalable storage solutions, preferably NetApp.</li>\n<li>Developing and maintaining performance metrics for systems.</li>\n<li>Strong understanding of Puppet Enterprise and Puppet Open Source.</li>\n<li>Strong understanding of basic networking concepts.</li>\n<li>Proficient with using Puppet to automate and manage configurations.</li>\n<li>Ability to translate business needs into technical solutions.</li>\n<li>Outstanding analytical and problem-solving skills.</li>\n<li>Strong written and verbal communication skills.</li>\n<li>Experience writing technical design documents and maintaining documentation.</li>\n<li>Familiarity with working in a source-controlled environment.</li>\n<li>Proficiency in scripting languages such as Bash, Python, or Perl.</li>\n<li>Experience with VMware services, including VMware data center virtualization and VMware Horizon.</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>Familiarity with DevOps practices and CI/CD tools.</li>\n<li>Knowledge of orchestration tools.</li>\n</ul>\n<p>Compensation and Benefits:</p>\n<p>$162,000 - $226,000</p>\n<p>Base salary is just one part of our total rewards package at xAI, which also includes equity, comprehensive medical, vision, and dental coverage, access to a 401(k) retirement plan, short &amp; long-term disability insurance, life insurance, and various other discounts and perks.</p>\n<p>ITAR Requirements:</p>\n<p>To conform to U.S. Government export regulations, applicant must be a (i) U.S. citizen or national, (ii) U.S. lawful, permanent resident (aka green card holder), (iii) Refugee under 8 U.S.C. § 1157, or (iv) Asylee under 8 U.S.C. § 1158, or be eligible to obtain the required authorizations from the U.S. Department of State. Learn more about the ITAR here.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e994a45c-88a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.xai.com/","logo":"https://logos.yubhub.co/xai.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/4871633007","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$162,000 - $226,000","x-skills-required":["infrastructure management","automation","systems engineering","corporate infrastructure","windows servers","linux servers","vmware technologies","puppet enterprise","configuration management","scalable storage solutions","netapp","performance metrics","basic networking concepts","scripting languages","bash","python","perl","vmware services","data center virtualization","horizon"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:38:18.741Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Palo Alto, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"infrastructure management, automation, systems engineering, corporate infrastructure, windows servers, linux servers, vmware technologies, puppet enterprise, configuration management, scalable storage solutions, netapp, performance metrics, basic networking concepts, scripting languages, bash, python, perl, vmware services, data center virtualization, horizon","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":162000,"maxValue":226000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a3e4b7f0-624"},"title":"Data Center Electrical Engineer","description":"<p>Training and serving frontier AI models requires compute infrastructure at a scale and density that pushes past what conventional data center designs were built to handle. As a Data Center Electrical Engineer, you&#39;ll own the electrical design of our facilities from the building service entrance through to the rack. You&#39;ll develop and maintain the reference designs and specifications our build partners work against, review their engineering submittals, and run the analysis needed to make confident decisions on topology, redundancy, and equipment selection.</p>\n<p>This is a highly cross-functional role. You&#39;ll work closely with our hardware and compute teams to translate accelerator requirements into electrical design criteria, and with supply chain to qualify equipment vendors and create optionality in a constrained market. Strong candidates will bring deep mission-critical electrical design experience and the judgment to make sound trade-offs when the standard playbook doesn&#39;t apply.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Develop and maintain Anthropic&#39;s electrical basis of design, reference architectures, and technical specifications for critical power distribution , covering switchgear, UPS systems, PDUs, busway, and rack power delivery</li>\n<li>Review and approve electrical design packages from build partners and MEP consultants, ensuring designs meet capacity, reliability, and maintainability requirements</li>\n<li>Perform and validate engineering studies including short-circuit, coordination, arc flash, load flow, and power quality analysis; use findings to steer design decisions and equipment selection</li>\n<li>Partner with external engineering and manufacturing teams to evaluate DC construction efficiency opportunities, develop productization roadmap and modular product strategy to enable accelerated DC acceleration and onsite labor reduction</li>\n<li>Evaluate and qualify electrical equipment vendors and product lines; work with supply chain to build a diversified vendor base that mitigates lead-time and single-source risk</li>\n<li>Drive standardization of electrical designs across sites to accelerate deployment timelines while preserving flexibility for site-specific constraints and evolving hardware generations</li>\n<li>Influence industry specifications and engage with equipment manufacturers on roadmap requirements for high-density, ML-optimized electrical infrastructure</li>\n<li>Provide technical oversight during construction and commissioning phases, resolving field engineering issues and reviewing test results against design intent</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Have 8+ years of electrical engineering experience in mission-critical facilities, with substantial time spent on data center or other high-availability electrical distribution design</li>\n<li>Hold a degree in Electrical Engineering or a closely related field; PE licensure is valued but not required</li>\n<li>Have hands-on experience producing and reviewing electrical construction documents, single-line diagrams, and equipment specifications, as well as development of prefabricated products and modularization towards offsite manufacturing</li>\n<li>Are fluent in power systems analysis and comfortable running or critically reviewing studies in tools such as SKM, ETAP, or EasyPower</li>\n<li>Understand critical power topologies (2N, N+1, distributed redundant, block redundant) and can articulate the trade-offs between them for different reliability and cost targets</li>\n<li>Have worked directly with electrical equipment manufacturers and can evaluate products on technical merit, not just datasheet claims</li>\n<li>Communicate clearly with both technical peers and non-specialist stakeholders, and can defend a design position when challenged</li>\n<li>Are results-oriented, with a bias toward practical solutions and a willingness to pick up work outside your core remit when the team needs it</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Experience building modular prefabricated products in electrical or mechanical areas of the datacenter</li>\n<li>Background in medium-voltage distribution, on-site generation, or energy storage integration</li>\n<li>Exposure to liquid-cooled infrastructure and its implications for electrical room layout and load distribution</li>\n<li>Familiarity with NFPA 70/70E, IEEE color book series, and relevant international electrical codes for multi-region deployments</li>\n<li>Prior work at a hyperscaler, colocation provider, or MEP consultancy serving large data center clients</li>\n<li>Experience driving reference designs or standards adopted across multiple sites or by external partners</li>\n</ul>\n<p>The annual compensation range for this role is $320,000-$405,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a3e4b7f0-624","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5187486008","x-work-arrangement":"remote-hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["Electrical engineering","Mission-critical facilities","Data center design","Power systems analysis","SKM, ETAP, or EasyPower","Critical power topologies","Electrical equipment manufacturers"],"x-skills-preferred":["Modular prefabricated products","Medium-voltage distribution","On-site generation","Energy storage integration","Liquid-cooled infrastructure","NFPA 70/70E, IEEE color book series"],"datePosted":"2026-04-18T15:37:22.035Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-Friendly, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Electrical engineering, Mission-critical facilities, Data center design, Power systems analysis, SKM, ETAP, or EasyPower, Critical power topologies, Electrical equipment manufacturers, Modular prefabricated products, Medium-voltage distribution, On-site generation, Energy storage integration, Liquid-cooled infrastructure, NFPA 70/70E, IEEE color book series","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e53014e6-57c"},"title":"Data Center Engineer, Resource Efficiency – Compute Supply","description":"<p>As a Power &amp; Resource Efficiency Engineer, you&#39;ll sit at the intersection of IT and facilities , building the systems, models, and control loops that optimize how we allocate and consume power, cooling, and physical capacity across our TPU/GPU fleet.</p>\n<p>You&#39;ll own the technical strategy for turning raw data center capacity into reliable, efficient compute, working across power topology, workload scheduling, and real-time telemetry to push utilization as close to the physical envelope as possible while maintaining our availability commitments.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Building models that forecast consumption across electrical and mechanical subsystems, informing capacity planning, energy procurement, oversubscription targets and risks, including statistical modeling of cluster utilization, workload profiles, and failure modes.</li>\n</ul>\n<ul>\n<li>Designing IT/OT interfaces that bridge compute orchestration with facility controls, enabling real-time telemetry across accelerator hardware, power distribution, cooling, and schedulers.</li>\n</ul>\n<ul>\n<li>Building and operating load management systems that use power and cooling topology to enable load management and power/thermal-aware placement to maximize throughput while meeting SLOs.</li>\n</ul>\n<ul>\n<li>Partnering with data center providers to drive design optimizations and hold them accountable to SLA-grade performance standards, providing technical diligence on partner architectures.</li>\n</ul>\n<p>In this role, you&#39;ll need to have deep knowledge of data center power distribution and cooling architectures, and how they interact with IT load profiles. Experience with reliability engineering, SLA development, and failure-mode analysis is also essential.</p>\n<p>Additionally, proficiency in statistical modeling and simulation for infrastructure capacity or power utilization, familiarity with SCADA/BMS/EPMS, telemetry pipelines, and control systems, and exposure to accelerator deployments and their power management interfaces are highly desirable.</p>\n<p>This is a challenging and rewarding role that requires a unique blend of technical expertise, business acumen, and collaboration skills. If you&#39;re passionate about data center infrastructure, AI, and sustainability, we encourage you to apply.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e53014e6-57c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5159642008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["data center power distribution and cooling architectures","_SYSTEMS","reliability engineering","SLA development","failure-mode analysis","statistical modeling and simulation","SCADA/BMS/EPMS","telemetry pipelines","control systems","accelerator deployments","power management interfaces"],"x-skills-preferred":["Python","similar languages","control theory","dynamical systems","cyber-physical systems design","energy storage","microgrid integration","demand response","behind-the-meter generation"],"datePosted":"2026-04-18T15:37:06.319Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-Friendly, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data center power distribution and cooling architectures, _SYSTEMS, reliability engineering, SLA development, failure-mode analysis, statistical modeling and simulation, SCADA/BMS/EPMS, telemetry pipelines, control systems, accelerator deployments, power management interfaces, Python, similar languages, control theory, dynamical systems, cyber-physical systems design, energy storage, microgrid integration, demand response, behind-the-meter generation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7783af60-46a"},"title":"Facilities Manager","description":"<p><strong>Job Description</strong></p>\n<p>As the Facilities Manager, you will lead a dynamic team dedicated to maintaining and enhancing our state-of-the-art infrastructure.</p>\n<p>Your role will be pivotal in ensuring seamless operations and creating an inspiring environment that reflects our commitment to excellence.</p>\n<p>By driving efficiency, managing critical projects, and fostering a culture of continuous improvement, you will help propel our mission of connecting people across the globe and supporting the future of life beyond Earth.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Lead the team responsible for all facilities operations (janitorial, landscaping, mechanics, HVAC, septic, water, trash, office supplies, build-outs, and general maintenance) at the facility to ensure zero unplanned interruptions of services to employees and customers</li>\n</ul>\n<ul>\n<li>Improve area safety and efficiency through regular auditing and continuous improvement</li>\n</ul>\n<ul>\n<li>Manage third-party service providers, such as janitorial teams, to ensure services are meeting corporate standards and terms of contract</li>\n</ul>\n<ul>\n<li>Develop, plan, and execute site improvement projects in accordance with overall business objectives and strategy</li>\n</ul>\n<ul>\n<li>Manage the maintenance, repair, and installation of all facilities related infrastructure such as buildings, HVAC equipment, septic, water infrastructure, and grounds</li>\n</ul>\n<ul>\n<li>Oversee facilities team, HVAC, and electrical technicians, drive overall performance and effectiveness through individual coaching and development</li>\n</ul>\n<ul>\n<li>Prioritize tasks and projects effectively to drive department efficiency and eliminate blockers to critical path</li>\n</ul>\n<p><strong>Basic Qualifications</strong></p>\n<ul>\n<li>High school diploma or equivalency certificate</li>\n</ul>\n<ul>\n<li>4+ years of professional experience in construction or facilities maintenance</li>\n</ul>\n<ul>\n<li>1+ years of leadership experience</li>\n</ul>\n<p><strong>Preferred Skills and Experience</strong></p>\n<ul>\n<li>Bachelor’s degree in construction management or an engineering discipline</li>\n</ul>\n<ul>\n<li>6+ years of professional experience in construction or facilities maintenance</li>\n</ul>\n<ul>\n<li>2+ years of leadership experience</li>\n</ul>\n<ul>\n<li>Proficiency in Microsoft Office Suite</li>\n</ul>\n<ul>\n<li>Experience managing budgets</li>\n</ul>\n<ul>\n<li>Experience managing third-party vendors</li>\n</ul>\n<ul>\n<li>Experience negotiating contracts and comparing service costs to achieve maximum value</li>\n</ul>\n<ul>\n<li>Experience with asset management software</li>\n</ul>\n<ul>\n<li>Knowledge of building management and automated facilities monitoring systems and processes: HVAC systems, fluid plumbing systems, fire suppression systems, generators, or fuel storage systems</li>\n</ul>\n<ul>\n<li>High attention to detail and cleanliness standards</li>\n</ul>\n<ul>\n<li>Ability to work independently with little direction</li>\n</ul>\n<ul>\n<li>Ability to meet deadlines in a fast-paced, ever-changing environment</li>\n</ul>\n<ul>\n<li>Ability to anticipate needs and develop innovative solutions for continuous improvement</li>\n</ul>\n<p><strong>Additional Requirements</strong></p>\n<ul>\n<li>Must be willing to work all shifts, extended hours, and weekends, as needed</li>\n</ul>\n<ul>\n<li>Schedule may vary depending on site operational needs and flexibility is required regarding overtime and weekend support</li>\n</ul>\n<ul>\n<li>Ability to perform job duties that require standing, kneeling, crouching, twisting upper body, and lifting up to 25 lbs. unassisted</li>\n</ul>\n<ul>\n<li>Must be willing to work onsite, remote work not considered</li>\n</ul>\n<p><strong>ITAR Requirements</strong></p>\n<ul>\n<li>To conform to U.S. Government export regulations, applicant must be a (i) U.S. citizen or national, (ii) U.S. lawful, permanent resident (aka green card holder), (iii) Refugee under 8 U.S.C. § 1157, or (iv) Asylee under 8 U.S.C. § 1158, or be eligible to obtain the required authorizations from the U.S. Department of State.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7783af60-46a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.xai.com/","logo":"https://logos.yubhub.co/xai.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/5080508007","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Microsoft Office Suite","Asset management software","Building management and automated facilities monitoring systems and processes","HVAC systems","Fluid plumbing systems","Fire suppression systems","Generators","Fuel storage systems"],"x-skills-preferred":["Construction management","Engineering discipline","Budget management","Third-party vendor management","Contract negotiation","Service cost comparison"],"datePosted":"2026-04-18T15:22:48.890Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Austin, TX"}},"employmentType":"FULL_TIME","occupationalCategory":"Operations","industry":"Technology","skills":"Microsoft Office Suite, Asset management software, Building management and automated facilities monitoring systems and processes, HVAC systems, Fluid plumbing systems, Fire suppression systems, Generators, Fuel storage systems, Construction management, Engineering discipline, Budget management, Third-party vendor management, Contract negotiation, Service cost comparison"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6f5337c-c2f"},"title":"Research Engineer (Scaling Multimodal Data)","description":"<p>We&#39;re looking for a research engineer to help improve our in-house world models through better multimodal data. This role is about figuring out what data actually moves model quality , then building the datasets, pipelines, and experiments to prove it.</p>\n<p>The best generative models aren’t just a product of model architecture and compute, they are a product of the training data. The model output reflects someone’s obsession over what goes into the data, how it’s processed, and what gets thrown away. We’re looking for the person who does the obsessing and builds the tools to act on it at scale.</p>\n<p>This isn’t a role where someone hands you a dataset and asks you to clean it. You will decide what data we need, figure out where to get it, build the processing and curation systems, and close the loop with model training to make sure it actually works.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Discover, evaluate, and acquire training data</li>\n<li>Build data processing and curation systems</li>\n<li>Look at the actual data constantly</li>\n<li>Close the data → model → evaluation loop</li>\n<li>Deploy ML models for data enrichment</li>\n<li>Make systematic, documented decisions</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>Strong software engineering fundamentals</li>\n<li>Deep experience with image and video data at scale</li>\n<li>Experience with distributed computing</li>\n<li>Experience using ML models as components</li>\n<li>A research-oriented approach to data decisions</li>\n<li>Familiarity with the model training lifecycle</li>\n</ul>\n<p><strong>Nice to Have:</strong></p>\n<ul>\n<li>Familiarity with columnar and large-scale data storage formats and libraries</li>\n<li>Track record of independently discovering and integrating new data sources into a training pipeline</li>\n<li>Direct experience closing the data → model quality loop</li>\n<li>Strong visual intuition for data quality and diversity</li>\n</ul>\n<p><strong>What This Isn’t:</strong></p>\n<ul>\n<li>Not infrastructure</li>\n<li>Not pure research</li>\n<li>Not a role where you wait for instructions</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6f5337c-c2f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"World Labs","sameAs":"https://world-labs.com/","logo":"https://logos.yubhub.co/world-labs.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/worldlabs/jobs/4164503009","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["software engineering fundamentals","image and video data at scale","distributed computing","ML models as components","research-oriented approach to data decisions","model training lifecycle"],"x-skills-preferred":["columnar and large-scale data storage formats and libraries","independently discovering and integrating new data sources","closing the data → model quality loop","visual intuition for data quality and diversity"],"datePosted":"2026-04-17T13:09:48.326Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"software engineering fundamentals, image and video data at scale, distributed computing, ML models as components, research-oriented approach to data decisions, model training lifecycle, columnar and large-scale data storage formats and libraries, independently discovering and integrating new data sources, closing the data → model quality loop, visual intuition for data quality and diversity"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6d4e84e5-9fa"},"title":"Pipeline Engineer (3D Data)","description":"<p>We&#39;re looking for a 3D Data Pipeline Engineer to design, build, and operate the core systems that enable high-quality 3D data processing, synthetic data generation, and rendering across our products.</p>\n<p>This is a hands-on role for someone who is passionate about large-scale 3D data, system performance, and delivering reliable data pipelines to power our product features.</p>\n<p>You&#39;ll work closely with product engineers, 3D artists, and research scientists to design efficient, robust, and scalable data pipeline capabilities,while keeping data integrity and performance high in a fast-moving startup environment.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Design, build, and operate automated pipelines for 3D data ingestion, cleaning, processing, validation, and delivery that sit on the critical path for model training.</li>\n<li>Own foundational capabilities for synthetic data generation, including developing tools, workflows, and quality metrics to produce high-fidelity training data at scale.</li>\n<li>Develop and optimize high-performance rendering systems and services for real-time visualization and asset generation.</li>\n<li>Architect and operate distributed data systems for handling massive volumes of 3D models, textures, and associated metadata, ensuring data consistency and robust failure recovery.</li>\n<li>Own data quality and production readiness end-to-end: defining data schemas, implementing quality checks, capacity planning, observability, and continuous improvement for the 3D pipeline.</li>\n<li>Improve developer and researcher velocity by building shared abstractions, tooling, and guardrails that reduce the operational and cognitive load of working with 3D assets.</li>\n<li>Collaborate with cross-functional teams to integrate the 3D data pipeline with other core product platforms and services.</li>\n<li>Set technical direction, mentor engineers, and raise the data engineering bar across the product org with a focus on 3D data.</li>\n</ul>\n<p><strong>Key Qualifications:</strong></p>\n<ul>\n<li>6+ years of experience building and operating large-scale data pipelines, especially with a focus on 3D, graphics, or simulation data, with deep experience designing scalable, distributed services in production.</li>\n<li>Strong programming skills in Python and/or C++, and a solid foundation in data engineering principles and distributed systems architecture.</li>\n<li>Hands-on experience with 3D data processing libraries, game engines (e.g., Unity, Unreal), or rendering APIs (e.g., OpenGL, Vulkan).</li>\n<li>Experience with cloud-based data storage and processing solutions (e.g., Kubernetes, distributed file systems, data warehouses).</li>\n<li>Experience working in fast-moving or startup environments, ideally having led systems or products from early design through production and growth.</li>\n<li>A high bar for ownership and execution: you’re comfortable with ambiguity, take responsibility for outcomes, and drive work forward without waiting for perfect clarity.</li>\n<li>A product-first mindset: you care about data quality, pipeline reliability, and performance as core product features, not afterthoughts.</li>\n<li>Enjoy collaborating with a small, high-ownership team and raising the quality bar through code, data design, and example.</li>\n</ul>\n<p><strong>Who You Are:</strong></p>\n<ul>\n<li>Fearless Innovator: We need people who thrive on challenges and aren&#39;t afraid to tackle the impossible.</li>\n<li>Resilient Builder: Impacting Large World Models isn&#39;t a sprint; it&#39;s a marathon with hurdles. We&#39;re looking for builders who can weather the storms of groundbreaking research and come out stronger.</li>\n<li>Mission-Driven Mindset: Everything we do is in service of creating the best spatially intelligent AI systems, and using them to empower people.</li>\n<li>Collaborative Spirit: We&#39;re building something bigger than any one person. We need team players who can harness the power of collective intelligence.</li>\n</ul>\n<p>We&#39;re hiring the brightest minds from around the globe to bring diverse perspectives to our cutting-edge work. If you&#39;re ready to work on technology that will reshape how machines perceive and interact with the world, World Labs is your launchpad.</p>\n<p>Join us, and let&#39;s make history together.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6d4e84e5-9fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"World Labs","sameAs":"https://www.worldlabs.ai","logo":"https://logos.yubhub.co/worldlabs.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/worldlabs/jobs/4110240009","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$250k-$350k base salary (good-faith estimate for San Francisco Bay Area upon hire; actual offer based on experience, skills, and qualifications)","x-skills-required":["Python","C++","3D data processing","Game engines (e.g., Unity, Unreal)","Rendering APIs (e.g., OpenGL, Vulkan)","Cloud-based data storage and processing solutions (e.g., Kubernetes, distributed file systems, data warehouses)"],"x-skills-preferred":[],"datePosted":"2026-04-17T13:09:44.260Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, C++, 3D data processing, Game engines (e.g., Unity, Unreal), Rendering APIs (e.g., OpenGL, Vulkan), Cloud-based data storage and processing solutions (e.g., Kubernetes, distributed file systems, data warehouses)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":250000,"maxValue":350000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6acd8036-5ec"},"title":"Platform Engineer (Databases & Storage)","description":"<p>We are looking for a Staff Platform Engineer to own the database and storage foundation of World Labs. This is a high-impact systems role at the intersection of databases, distributed systems, and AI infrastructure. You will define how core data systems are designed, scaled, and operated in an environment where workloads are evolving quickly and requirements are often ambiguous.</p>\n<p>Your responsibilities will include owning the design and evolution of the transactional systems that power the platform, defining architecture for database and storage systems under high-throughput, low-latency workloads, making and driving decisions around data modeling, indexing, replication, and consistency, debugging and resolving complex production issues, establishing standards for reliability, observability, and operability across the platform, partnering with product and research teams to support evolving and often ambiguous requirements, driving improvements in performance, scalability, and cost across the system, mentoring engineers and raising the bar for system design and technical decision-making.</p>\n<p>Key qualifications include 10+ years of experience building and operating production systems at scale, with ownership of critical infrastructure, strong experience designing and operating transactional systems and databases, deep understanding of data modeling, indexing, transactions, concurrency, and consistency tradeoffs, experience owning systems with strict reliability and performance requirements in production, strong experience debugging complex production issues and reasoning about failure modes, experience designing distributed systems or large-scale infrastructure where tradeoffs are non-trivial, proven ability to define architecture and drive technical decisions end-to-end, strong judgment in balancing performance, reliability, and cost, ability to operate effectively in ambiguous, fast-moving environments with high ownership.</p>\n<p>Preferred qualifications include experience with database internals, storage systems, or query engines, experience building infrastructure for AI/ML systems or data platforms, experience in early-stage or high-growth environments.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6acd8036-5ec","directApply":true,"hiringOrganization":{"@type":"Organization","name":"World Labs","sameAs":"https://www.worldlabs.ai","logo":"https://logos.yubhub.co/worldlabs.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/worldlabs/jobs/4194381009","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$200-$300k base salary (good-faith estimate for San Francisco Bay Area upon hire; actual offer based on experience, skills, and qualifications)","x-skills-required":["database internals","storage systems","query engines","data modeling","indexing","transactions","concurrency","consistency","distributed systems","large-scale infrastructure"],"x-skills-preferred":["AI/ML systems","data platforms"],"datePosted":"2026-04-17T13:09:33.493Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database internals, storage systems, query engines, data modeling, indexing, transactions, concurrency, consistency, distributed systems, large-scale infrastructure, AI/ML systems, data platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":300000,"maxValue":300000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_de4049d7-242"},"title":"Senior Electrical Engineer","description":"<p>Saronic Technologies is seeking a Senior Electrical Engineer specializing in ruggedized computing and networking systems to join our Electrical Engineering – Advanced Development group.</p>\n<p>This role will focus on the research, evaluation, and development of robust onboard computing architectures, embedded systems, and high-reliability network solutions that power Saronic’s autonomous vessel platforms.</p>\n<p>The ideal candidate will have deep experience designing and validating ruggedized electronic systems for mission-critical applications, including embedded computing, network topologies, data management, and environmental hardening for commercial maritime and defense use cases.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li><p>Lead R&amp;D initiatives in ruggedized computing and networking architectures for autonomous surface vessels.</p>\n</li>\n<li><p>Design, evaluate, and integrate embedded computing systems, data acquisition units, and network infrastructures optimized for high performance in harsh marine environments.</p>\n</li>\n<li><p>Conduct benchmarking and trade studies on ruggedized COTS and custom computing solutions (edge computers, network switches, routers, storage units, etc.).</p>\n</li>\n<li><p>Develop and validate system architectures for high-availability networks supporting autonomy, sensing, and control subsystems.</p>\n</li>\n<li><p>Collaborate with software, autonomy, and mechanical engineering teams to ensure reliable data throughput and system resilience across vessel networks.</p>\n</li>\n<li><p>Specify and validate environmental and EMC/EMI compliance for computing and networking hardware.</p>\n</li>\n<li><p>Prototype and test system configurations in laboratory and field conditions, including shock, vibration, temperature, and humidity testing.</p>\n</li>\n<li><p>Author technical documentation, including R&amp;D reports, trade studies, wiring diagrams, and integration standards.</p>\n</li>\n<li><p>Mentor junior engineers and contribute to internal design guidelines for next-generation computing and network systems.</p>\n</li>\n<li><p>Support system integration and troubleshooting during prototype builds, dockside commissioning, and sea trials.</p>\n</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li><p>B.S. or M.S. in Electrical Engineering, Computer Engineering, or related discipline.</p>\n</li>\n<li><p>7+ years of experience in electrical or systems engineering focused on computing and networking technologies in ruggedized or mission-critical environments.</p>\n</li>\n<li><p>Expertise in embedded computing platforms, network design, and hardware integration.</p>\n</li>\n<li><p>Experience with Ethernet, CAN, serial, and fiber-optic communication protocols and their implementation in distributed systems.</p>\n</li>\n<li><p>Proven track record of benchmarking and trade study development for hardware performance and reliability.</p>\n</li>\n<li><p>Familiarity with marine, aerospace, automotive or defense environmental standards (MIL-STD-810, MIL-STD-461, IEC 60945, etc.).</p>\n</li>\n<li><p>Strong understanding of power distribution, grounding, and thermal management in dense electronics enclosures.</p>\n</li>\n<li><p>Excellent communication skills and experience producing clear technical documentation and reports.</p>\n</li>\n<li><p>Hands-on experience with system integration and environmental testing.</p>\n</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li><p>Experience developing or integrating ruggedized computing solutions for maritime or defense systems.</p>\n</li>\n<li><p>Familiarity with network security, IEEE 1588/PTP Protocol, VLAN management, and deterministic networking for real-time systems.</p>\n</li>\n<li><p>Knowledge of data logging, storage, and redundancy architectures in distributed sensor networks.</p>\n</li>\n<li><p>Experience with hardware-in-the-loop (HITL) and hardware-software co-simulation environments.</p>\n</li>\n<li><p>Background in autonomous or remote vehicle platforms.</p>\n</li>\n<li><p>Understanding of cybersecurity standards and secure network design principles.</p>\n</li>\n<li><p>Experience using 3D CAD tools to communicate with other engineering groups (e.g. Siemens NX, Creo, SolidWorks)</p>\n</li>\n<li><p>Experience utilizing ECAD tools to define/draw single line diagrams and schematics (e.g. Altium, Zuken, AutoCAD, Siemens Capital)</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_de4049d7-242","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Saronic Technologies","sameAs":"https://www.saronictechnologies.com/","logo":"https://logos.yubhub.co/saronictechnologies.com.png"},"x-apply-url":"https://jobs.lever.co/saronic/ade089f5-be71-4d84-bf7d-2ba931fce248","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["embedded computing platforms","network design","hardware integration","Ethernet","CAN","serial","fiber-optic communication protocols","distributed systems","benchmarking","trade study development","hardware performance and reliability","marine","aerospace","automotive","defense environmental standards","power distribution","grounding","thermal management","dense electronics enclosures","communication skills","technical documentation","system integration","environmental testing"],"x-skills-preferred":["ruggedized computing solutions","maritime or defense systems","network security","IEEE 1588/PTP Protocol","VLAN management","deterministic networking","real-time systems","data logging","storage","redundancy architectures","distributed sensor networks","hardware-in-the-loop","hardware-software co-simulation environments","autonomous or remote vehicle platforms","cybersecurity standards","secure network design principles","3D CAD tools","ECAD tools","single line diagrams","schematics"],"datePosted":"2026-04-17T12:58:22.001Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"embedded computing platforms, network design, hardware integration, Ethernet, CAN, serial, fiber-optic communication protocols, distributed systems, benchmarking, trade study development, hardware performance and reliability, marine, aerospace, automotive, defense environmental standards, power distribution, grounding, thermal management, dense electronics enclosures, communication skills, technical documentation, system integration, environmental testing, ruggedized computing solutions, maritime or defense systems, network security, IEEE 1588/PTP Protocol, VLAN management, deterministic networking, real-time systems, data logging, storage, redundancy architectures, distributed sensor networks, hardware-in-the-loop, hardware-software co-simulation environments, autonomous or remote vehicle platforms, cybersecurity standards, secure network design principles, 3D CAD tools, ECAD tools, single line diagrams, schematics"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c629a0da-f6c"},"title":"Security Engineer","description":"<p>We&#39;re seeking a Security Engineer at the senior-level or above focused on hardware, embedded systems, and firmware security to own the security posture of Saronic&#39;s vessel hardware platforms from silicon to system.</p>\n<p>You will be the technical authority on hardware root of trust, secure boot, firmware integrity, embedded system hardening, and the security of third-party hardware integrations. Your work ensures that every component on the vessel is resilient against tampering, exploitation, and supply chain compromise, designed in from the start and maintained across the fleet lifecycle.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li><p>Conduct hardware security assessments including fault injection, side-channel analysis, interface evaluation, and bus protocol analysis across Saronic-built and third-party hardware including sensors, radios, navigation systems, propulsion controllers, and communication modules</p>\n</li>\n<li><p>Evaluate and harden physical interfaces, debug ports, maintenance access points, and removable media interfaces on vessel hardware</p>\n</li>\n<li><p>Evaluate supply chain security risks for hardware components and recommend provenance validation, anti-tamper, and attestation controls</p>\n</li>\n<li><p>Develop and maintain a hardware security testing capability including tooling, methodology, and repeatable test procedures</p>\n</li>\n<li><p>Design and implement secure boot chains establishing hardware root of trust from power-on through application launch, integrating TPM, secure elements, and HSMs for device identity, key storage, measured boot, and remote attestation</p>\n</li>\n<li><p>Design and implement secure firmware update mechanisms including signed updates, rollback protection, and verified delivery across the fleet</p>\n</li>\n<li><p>Own the cryptographic key lifecycle for hardware-bound keys, including provisioning, rotation, revocation, and escrow</p>\n</li>\n<li><p>Harden embedded Linux systems on vessel platforms, including kernel configuration, mandatory access controls, secure IPC, and attack surface reduction</p>\n</li>\n<li><p>Secure operational technology protocols and interfaces used in vessel control systems, propulsion, navigation, and sensor fusion including CAN bus, NMEA, and maritime/industrial communication protocols</p>\n</li>\n<li><p>Define security boundaries, trust zones, and segmentation strategies for vessel-internal compute and communication architectures</p>\n</li>\n<li><p>Drive threat modeling across vessel hardware subsystems and translate findings into actionable engineering requirements</p>\n</li>\n<li><p>Produce secure-by-design reference architectures and define hardware and firmware security standards, testing requirements, and acceptance criteria integrated into engineering workflows</p>\n</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li><p>6+ years of hands-on experience in hardware security, embedded systems security, firmware security, or a closely related security engineering role</p>\n</li>\n<li><p>Deep expertise in hardware hacking techniques including fault injection, side-channel attacks, JTAG/SWD exploitation, bus sniffing/injection, and physical security assessments</p>\n</li>\n<li><p>Demonstrated experience designing and implementing secure boot chains, hardware root of trust, and secure firmware update mechanisms in production systems</p>\n</li>\n<li><p>Strong experience assessing third-party hardware integrations and evaluating supply chain security risks</p>\n</li>\n<li><p>Deep knowledge of embedded Linux security hardening, kernel security, and mandatory access control frameworks</p>\n</li>\n<li><p>Experience with operational technology security, industrial protocols, or control system security</p>\n</li>\n<li><p>Proficiency in C, C++, Python, or Rust in the context of firmware, embedded, or systems-level security work, and with hardware security testing tools</p>\n</li>\n<li><p>Ability to obtain and maintain a security clearance</p>\n</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li><p>Experience in defense, aerospace, robotics, autonomy, maritime, or other high-assurance environments</p>\n</li>\n<li><p>Experience with autonomous systems, unmanned vehicles, or safety-critical embedded platforms</p>\n</li>\n<li><p>Experience with RTOS, microcontroller security, or resource-constrained device environments</p>\n</li>\n<li><p>Knowledge of CAN bus, NMEA protocols, maritime communication systems, RF/GPS/GNSS security, or ICS security standards</p>\n</li>\n<li><p>Familiarity with defense or safety-critical compliance frameworks (NIST SP 800-53, IEC 62443, Common Criteria, or equivalent)</p>\n</li>\n<li><p>Relevant certifications such as OSEE, GXPN, GSE, or hardware-focused credentials</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c629a0da-f6c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Saronic Technologies","sameAs":"https://www.saronictechnologies.com/","logo":"https://logos.yubhub.co/saronictechnologies.com.png"},"x-apply-url":"https://jobs.lever.co/saronic/4b15b1b4-3c34-47ad-b964-dbcf0f8a3dc4","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Hardware security","Embedded systems security","Firmware security","Fault injection","Side-channel analysis","Interface evaluation","Bus protocol analysis","Physical security assessments","Secure boot chains","Hardware root of trust","Firmware integrity","Embedded system hardening","Third-party hardware integrations","Supply chain security risks","Provenance validation","Anti-tamper","Attestation controls","Hardware security testing","Tooling","Methodology","Repeatable test procedures","Device identity","Key storage","Measured boot","Remote attestation","Signed updates","Rollback protection","Verified delivery","Cryptographic key lifecycle","Provisioning","Rotation","Revocation","Escrow","Embedded Linux systems","Kernel configuration","Mandatory access controls","Secure IPC","Attack surface reduction","Operational technology protocols","Industrial protocols","Control system security","CAN bus","NMEA","Maritime/industrial communication protocols","Security boundaries","Trust zones","Segmentation strategies","Threat modeling","Actionable engineering requirements","Secure-by-design reference architectures","Hardware and firmware security standards","Testing requirements","Acceptance criteria","Engineering workflows","C","C++","Python","Rust","Hardware security testing tools"],"x-skills-preferred":["Defense","Aerospace","Robotics","Autonomy","Maritime","High-assurance environments","Autonomous systems","Unmanned vehicles","Safety-critical embedded platforms","RTOS","Microcontroller security","Resource-constrained device environments","NMEA protocols","Maritime communication systems","RF/GPS/GNSS security","ICS security standards","Defense or safety-critical compliance frameworks","OSEE","GXPN","GSE","Hardware-focused credentials"],"datePosted":"2026-04-17T12:57:49.070Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Hardware security, Embedded systems security, Firmware security, Fault injection, Side-channel analysis, Interface evaluation, Bus protocol analysis, Physical security assessments, Secure boot chains, Hardware root of trust, Firmware integrity, Embedded system hardening, Third-party hardware integrations, Supply chain security risks, Provenance validation, Anti-tamper, Attestation controls, Hardware security testing, Tooling, Methodology, Repeatable test procedures, Device identity, Key storage, Measured boot, Remote attestation, Signed updates, Rollback protection, Verified delivery, Cryptographic key lifecycle, Provisioning, Rotation, Revocation, Escrow, Embedded Linux systems, Kernel configuration, Mandatory access controls, Secure IPC, Attack surface reduction, Operational technology protocols, Industrial protocols, Control system security, CAN bus, NMEA, Maritime/industrial communication protocols, Security boundaries, Trust zones, Segmentation strategies, Threat modeling, Actionable engineering requirements, Secure-by-design reference architectures, Hardware and firmware security standards, Testing requirements, Acceptance criteria, Engineering workflows, C, C++, Python, Rust, Hardware security testing tools, Defense, Aerospace, Robotics, Autonomy, Maritime, High-assurance environments, Autonomous systems, Unmanned vehicles, Safety-critical embedded platforms, RTOS, Microcontroller security, Resource-constrained device environments, NMEA protocols, Maritime communication systems, RF/GPS/GNSS security, ICS security standards, Defense or safety-critical compliance frameworks, OSEE, GXPN, GSE, Hardware-focused credentials"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f7ce6c3c-a15"},"title":"Senior Software Engineer — Admin Platform","description":"<p>As a Senior Software Engineer on the XCore team, you will work on one of the most impactful projects at People.ai , Forecasting. You will be building and evolving a live product with active daily users, working closely with the business team and customers through discovery and exploration calls. You will have a direct impact on company revenue by defining and delivering features grounded in real customer feedback. Your work will directly influence how customers plan, forecast outcomes, and make critical business decisions. In addition, you will collaborate across Engineering and R&amp;D teams to enhance core Admin Platform services such as authentication, users, settings, and configurations. You will leverage customer insights and data analysis to design and build high-performance, low-latency systems and APIs that scale efficiently while improving our operational costs</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Take ownership of projects. From requirements gathering, through design, implementation and testing, to monitoring and maintenance throughout the tech stack.</li>\n<li>Collaborate within the team, with other engineering teams and cross-functionally to enhance our users’ experience.</li>\n<li>Follow software design and development best practices and promote such practices in the team.</li>\n<li>Document design choices and operational knowledge to successfully deploy and run services.</li>\n<li>Ensure code accuracy and efficiency with appropriate test coverage, integration testing and load testing.</li>\n<li>Ensure robust alerting, dashboards, and runbooks for production services are in place.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>7+ years of professional experience working on backend systems in an enterprise environment</li>\n<li>2+ years experience programming in Python 2.x/3.x or Scala or Java</li>\n<li>Experience with SOA, microservices, and event-driven architecture</li>\n<li>Experience with an enterprise-grade stack for scalable web apps, including messaging brokers, in-memory storage, and key-value databases</li>\n<li>Strong knowledge of TDD, Unit, and automated test paradigms</li>\n<li>Experience with SQL and RDBMS solutions</li>\n<li>Experience with containerized applications, Docker, and Kubernetes</li>\n<li>Possess a DevOps mindset, AWS experience is a plus</li>\n<li>Experience with LLMs is a plus</li>\n<li>Experience with large-scale data processing (Spark) is a plus</li>\n<li>Bachelor’s Degree in Computer Science, Computer Engineering, or in a closely related discipline</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f7ce6c3c-a15","directApply":true,"hiringOrganization":{"@type":"Organization","name":"People.ai","sameAs":"https://www.people.ai","logo":"https://logos.yubhub.co/people.ai.png"},"x-apply-url":"https://jobs.lever.co/people-ai/a52ff2af-7ee1-4afb-a438-71939081b44f","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"contract","x-salary-range":null,"x-skills-required":["Python","Java","Scala","SOA","Microservices","Event-driven architecture","Enterprise-grade stack","Messaging brokers","In-memory storage","Key-value databases","TDD","Unit","Automated test paradigms","SQL","RDBMS","Containerized applications","Docker","Kubernetes","DevOps","AWS"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:52:27.117Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Poland"}},"jobLocationType":"TELECOMMUTE","employmentType":"CONTRACTOR","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, SOA, Microservices, Event-driven architecture, Enterprise-grade stack, Messaging brokers, In-memory storage, Key-value databases, TDD, Unit, Automated test paradigms, SQL, RDBMS, Containerized applications, Docker, Kubernetes, DevOps, AWS"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dfb22cd6-11a"},"title":"Staff Software Engineer - Online Storage","description":"<p>We believe that the way people interact with their finances will drastically improve in the next few years. We&#39;re dedicated to empowering this transformation by building the tools and experiences that thousands of developers use to create their own products. The Online Storage team is growing! We build the storage layer backing Plaid&#39;s Core Data Models used by all of engineering. The goal is to evolve Plaid&#39;s storage systems to be more reliable, scalable, performant, and efficient.</p>\n<p>As an engineer on the storage team, you will have the opportunity to influence the strategy of the team and design, deploy, and scale the online storage platform at Plaid. You will build services and tools to improve operation automations. You will optimize query performance and infrastructure cost. You will wield terraform and cloud resources with confidence and certainty. You will partner with our stakeholder teams to understand their current storage and future storage requirements. You will improve the user experience that partner teams have interacting with the storage platform.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build and scale systems to support large datasets and high-throughput workloads with strong performance and reliability guarantees</li>\n<li>Design and evolve resilient, distributed storage systems that operate reliably</li>\n<li>Navigate ambiguity and drive clarity across complex storage and infrastructure challenges</li>\n<li>Diagnose and resolve deeply technical issues within storage engines, including analyzing internal behavior and, when necessary, diving into source code</li>\n<li>Lead careful, reliability-focused migrations that minimize risk and ensure seamless transitions for dependent systems</li>\n<li>Build foundational storage capabilities that significantly reduce friction for product engineers, enabling teams to move faster and more confidently</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>Strong software engineering experience with deep focus on backend and/or platform systems</li>\n<li>Deep expertise in storage technologies</li>\n<li>Experience deploying and orchestrating databases at scale</li>\n<li>Strong background in database performance optimization, preferably with SQL databases</li>\n<li>Experience designing and evolving storage platforms</li>\n<li>Infrastructure experience with tools such as Terraform and AWS</li>\n<li>Hands-on experience with databases such as TiDB, MySQL, or MongoDB</li>\n<li>Experience building internal tooling and automation to improve reliability and developer efficiency</li>\n</ul>\n<p><strong>Nice to Have</strong></p>\n<ul>\n<li>Demonstrated curiosity about evolving storage technologies and performance optimization techniques</li>\n<li>Production experience with Go and TypeScript</li>\n<li>Experience applying security best practices in storage or infrastructure systems</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dfb22cd6-11a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Plaid","sameAs":"https://plaid.com/","logo":"https://logos.yubhub.co/plaid.com.png"},"x-apply-url":"https://jobs.lever.co/plaid/90b7be5b-30f3-43c0-b576-879065991072","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$207,600-$339,600 per year","x-skills-required":["storage technologies","database performance optimization","SQL databases","Terraform","AWS","TiDB","MySQL","MongoDB"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:52:22.974Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"storage technologies, database performance optimization, SQL databases, Terraform, AWS, TiDB, MySQL, MongoDB","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":207600,"maxValue":339600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b1a53764-a49"},"title":"Senior Software Engineer — Admin Platform","description":"<p>As a Senior Software Engineer on the XCore team, you will work on one of the most impactful projects at People.ai , Forecasting. You will be building and evolving a live product with active daily users, working closely with the business team and customers through discovery and exploration calls. You will have a direct impact on company revenue by defining and delivering features grounded in real customer feedback. Your work will directly influence how customers plan, forecast outcomes, and make critical business decisions.</p>\n<p>In addition, you will collaborate across Engineering and R&amp;D teams to enhance core Admin Platform services such as authentication, users, settings, and configurations. You will leverage customer insights and data analysis to design and build high-performance, low-latency systems and APIs that scale efficiently while improving our operational costs.</p>\n<p>We value ownership highly , the ability to take an idea through all the stages from conception to shipping a product. This reflects throughout our company, but is especially true in engineering. As an engineer at People.ai, you&#39;ll be a part of a highly independent and autonomous team.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Take ownership of projects. From requirements gathering, through design, implementation and testing, to monitoring and maintenance throughout the tech stack.</li>\n<li>Collaborate within the team, with other engineering teams and cross-functionally to enhance our users’ experience.</li>\n<li>Follow software design and development best practices and promote such practices in the team.</li>\n<li>Document design choices and operational knowledge to successfully deploy and run services.</li>\n<li>Ensure code accuracy and efficiency with appropriate test coverage, integration testing and load testing.</li>\n<li>Ensure robust alerting, dashboards, and runbooks for production services are in place.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>7+ years of professional experience working on backend systems in an enterprise environment</li>\n<li>2+ years experience programming in Python 2.x/3.x or Scala or Java</li>\n<li>Experience with SOA, microservices, and event-driven architecture</li>\n<li>Experience with an enterprise-grade stack for scalable web apps, including messaging brokers, in-memory storage, and key-value databases</li>\n<li>Strong knowledge of TDD, Unit, and automated test paradigms</li>\n<li>Experience with SQL and RDBMS solutions</li>\n<li>Experience with containerized applications, Docker, and Kubernetes</li>\n<li>Possess a DevOps mindset, AWS experience is a plus</li>\n<li>Experience with LLMs is a plus</li>\n<li>Experience with large-scale data processing (Spark) is a plus</li>\n<li>Bachelor’s Degree in Computer Science, Computer Engineering, or in a closely related discipline</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b1a53764-a49","directApply":true,"hiringOrganization":{"@type":"Organization","name":"People.ai","sameAs":"https://www.people.ai","logo":"https://logos.yubhub.co/people.ai.png"},"x-apply-url":"https://jobs.lever.co/people-ai/111fc468-37d0-4b08-8576-746a96f0ec85","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"contract","x-salary-range":null,"x-skills-required":["Python","Java","Scala","SOA","Microservices","Event-driven architecture","Enterprise-grade stack","Messaging brokers","In-memory storage","Key-value databases","TDD","Unit testing","Automated testing","SQL","RDBMS","Containerized applications","Docker","Kubernetes","DevOps","AWS","LLMs","Large-scale data processing","Spark"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:52:17.777Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Prague"}},"jobLocationType":"TELECOMMUTE","employmentType":"CONTRACTOR","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, SOA, Microservices, Event-driven architecture, Enterprise-grade stack, Messaging brokers, In-memory storage, Key-value databases, TDD, Unit testing, Automated testing, SQL, RDBMS, Containerized applications, Docker, Kubernetes, DevOps, AWS, LLMs, Large-scale data processing, Spark"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9d3a998b-29b"},"title":"Senior Software Engineer — Admin Platform","description":"<p>As a Senior Software Engineer on the XCore team, you will work on one of the most impactful projects at People.ai , Forecasting. You will be building and evolving a live product with active daily users, working closely with the business team and customers through discovery and exploration calls. You will have a direct impact on company revenue by defining and delivering features grounded in real customer feedback. Your work will directly influence how customers plan, forecast outcomes, and make critical business decisions.</p>\n<p>In addition, you will collaborate across Engineering and R&amp;D teams to enhance core Admin Platform services such as authentication, users, settings, and configurations. You will leverage customer insights and data analysis to design and build high-performance, low-latency systems and APIs that scale efficiently while improving our operational costs.</p>\n<p>We value ownership highly , the ability to take an idea through all the stages from conception to shipping a product. This reflects throughout our company, but is especially true in engineering. As an engineer at People.ai, you&#39;ll be a part of a highly independent and autonomous team.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Take ownership of projects. From requirements gathering, through design, implementation and testing, to monitoring and maintenance throughout the tech stack.</li>\n<li>Collaborate within the team, with other engineering teams and cross-functionally to enhance our users’ experience.</li>\n<li>Follow software design and development best practices and promote such practices in the team.</li>\n<li>Document design choices and operational knowledge to successfully deploy and run services.</li>\n<li>Ensure code accuracy and efficiency with appropriate test coverage, integration testing and load testing.</li>\n<li>Ensure robust alerting, dashboards, and runbooks for production services are in place.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>7+ years of professional experience working on backend systems in an enterprise environment</li>\n<li>2+ years experience programming in Python 2.x/3.x or Scala or Java</li>\n<li>Experience with SOA, microservices, and event-driven architecture</li>\n<li>Experience with an enterprise-grade stack for scalable web apps, including messaging brokers, in-memory storage, and key-value databases</li>\n<li>Strong knowledge of TDD, Unit, and automated test paradigms</li>\n<li>Experience with SQL and RDBMS solutions</li>\n<li>Experience with containerized applications, Docker, and Kubernetes</li>\n<li>Possess a DevOps mindset, AWS experience is a plus</li>\n<li>Experience with LLMs is a plus</li>\n<li>Experience with large-scale data processing (Spark) is a plus</li>\n<li>Bachelor’s Degree in Computer Science, Computer Engineering, or in a closely related discipline</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9d3a998b-29b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"People.ai","sameAs":"https://www.people.ai","logo":"https://logos.yubhub.co/people.ai.png"},"x-apply-url":"https://jobs.lever.co/people-ai/01b49f33-5471-4459-878a-6d69749b2671","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"contract","x-salary-range":null,"x-skills-required":["Python","Java","Scala","SOA","Microservices","Event-driven architecture","Enterprise-grade stack","Messaging brokers","In-memory storage","Key-value databases","TDD","Unit testing","Automated testing","SQL","RDBMS","Containerized applications","Docker","Kubernetes","DevOps","AWS","LLMs","Large-scale data processing","Spark"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:51:55.889Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Romania"}},"jobLocationType":"TELECOMMUTE","employmentType":"CONTRACTOR","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, SOA, Microservices, Event-driven architecture, Enterprise-grade stack, Messaging brokers, In-memory storage, Key-value databases, TDD, Unit testing, Automated testing, SQL, RDBMS, Containerized applications, Docker, Kubernetes, DevOps, AWS, LLMs, Large-scale data processing, Spark"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dd7fb909-289"},"title":"Web Crawling Engineer","description":"<p>About Mistral AI</p>\n<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>\n<p>We are looking for a skilled and motivated Web Crawling Engineer to join our dynamic engineering team. The ideal candidate should have a solid background in distributed web crawling, scraping and data extraction, with experience using advanced tools and technologies to collect and process large-scale data from diverse web sources at large scale.</p>\n<p>Responsibilities</p>\n<p>As a Web crawling engineer, you will be responsible for:</p>\n<ul>\n<li>Developing and maintaining web crawlers using Go to extract data from target websites.</li>\n<li>Utilizing headless browsing techniques, such as Chrome DevTools, to automate and optimize data collection processes.</li>\n<li>Collaborating with cross-functional teams to identify, scrape, and integrate data from APIs and web pages to support business objectives.</li>\n<li>Creating and implementing efficient parsing patterns using tokenizers, regular expressions, XPaths, and CSS selectors to ensure accurate data extraction.</li>\n<li>Designing and managing distributed job queues using technologies such as Redis, Aerospike and Kubernetes to handle large-scale distributed crawling and processing tasks.</li>\n<li>Developing strategies to monitor and ensure data quality, accuracy, and integrity throughout the crawling and indexing process.</li>\n<li>Continuously improving and optimizing existing web crawling infrastructure to maximize efficiency and adapt to new challenges.</li>\n</ul>\n<p>About You</p>\n<p>Core programming and web technologies</p>\n<ul>\n<li>Proficiency in Go (Golang)/Rust/Zig for building scalable and efficient web crawlers.</li>\n<li>Deep understanding of TCP, UDP, TLS and HTTP/1.1,2,3 protocols and web communication.</li>\n<li>Knowledge of HTML, CSS, and JavaScript for parsing and navigating web content.</li>\n<li>Familiarity with cloud platforms (AWS, GCP), orchestration (Kubernetes, Nomad), and containerization (Docker) for deployment.</li>\n</ul>\n<p>Data Structures &amp; Algorithms</p>\n<ul>\n<li>Mastery of queues, stacks, hash maps, and other data structures for efficient data handling.</li>\n<li>Ability to design and optimize algorithms for large-scale web crawling.</li>\n</ul>\n<p>Web Scraping &amp; Data Acquisition</p>\n<ul>\n<li>Hands-on experience with networking and web scraping libraries.</li>\n<li>Understanding of how search engines work and best practices for web crawling optimization.</li>\n</ul>\n<p>Databases &amp; Data Storage</p>\n<ul>\n<li>Experience with SQL and/or NoSQL databases (knowing Aerospike is a bonus) for storing and managing crawled data.</li>\n<li>Familiarity with data warehousing and scalable storage solutions.</li>\n</ul>\n<p>Distributed Systems &amp; Big Data</p>\n<ul>\n<li>Knowledge of distributed systems (e.g., Hadoop, Spark) for processing large datasets.</li>\n</ul>\n<p>Bonus Skills (Nice-to-Have)</p>\n<ul>\n<li>Experience with web archiving projects &amp; tooling, open-source archiving is a big plus!</li>\n<li>Experience applying Machine Learning to improve crawling efficiency or accuracy.</li>\n<li>Experience with low-level networking programming and/or userspace TCP/IP stacks.</li>\n</ul>\n<p>Hiring Process</p>\n<p>Here is what you should expect:</p>\n<ul>\n<li>Introduction call - 35 min</li>\n<li>Hiring Manager Interview - 30 min</li>\n<li>Live-coding Interview - 45 min</li>\n<li>System Design Interview - 45 min</li>\n<li>Deep dive interview (optional) - 60min</li>\n<li>Culture-fit discussion - 30 min</li>\n<li>Reference checks</li>\n</ul>\n<p>Additional Information</p>\n<p>Location &amp; Remote</p>\n<p>This role is primarily based in one of our European offices , Paris, France and London, UK. We will prioritize candidates who either reside there or are open to relocating. We strongly believe in the value of in-person collaboration to foster strong relationships and seamless communication within our team. In certain specific situations, we will also consider remote candidates based in one of the countries listed in this job posting , currently France, UK, Germany, Belgium, Netherlands, Spain and Italy. In any case, we ask all new hires to visit our Paris HQ office:</p>\n<ul>\n<li>for the first week of their onboarding (accommodation and travelling covered)</li>\n<li>then at least 2 days per month</li>\n</ul>\n<p>What we offer</p>\n<p>💰 Competitive salary and equity</p>\n<p>🧑‍⚕️ Health insurance</p>\n<p>🚴 Transportation allowance</p>\n<p>🥎 Sport allowance</p>\n<p>🥕 Meal vouchers</p>\n<p>💰 Private pension plan</p>\n<p>🍼 Parental : Generous parental leave policy</p>\n<p>🌎 Visa sponsorship</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dd7fb909-289","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Mistral AI","sameAs":"https://mistral.ai","logo":"https://logos.yubhub.co/mistral.ai.png"},"x-apply-url":"https://jobs.lever.co/mistral/c96bf665-7d73-406b-8d8f-ddf8df5d160f","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Rust","Zig","TCP","UDP","TLS","HTTP/1.1","HTTP/2","HTTP/3","HTML","CSS","JavaScript","cloud platforms","orchestration","containerization","queues","stacks","hash maps","SQL","NoSQL databases","data warehousing","scalable storage solutions","distributed systems","Hadoop","Spark"],"x-skills-preferred":["web archiving projects","Machine Learning","low-level networking programming","userspace TCP/IP stacks"],"datePosted":"2026-04-17T12:48:06.790Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Paris"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Rust, Zig, TCP, UDP, TLS, HTTP/1.1, HTTP/2, HTTP/3, HTML, CSS, JavaScript, cloud platforms, orchestration, containerization, queues, stacks, hash maps, SQL, NoSQL databases, data warehousing, scalable storage solutions, distributed systems, Hadoop, Spark, web archiving projects, Machine Learning, low-level networking programming, userspace TCP/IP stacks"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6f25b435-69f"},"title":"Technical Support Engineer – On-Premise","description":"<p>We are seeking a Technical Support Engineer - On-Premise Infrastructure to join our Support team in France. This role is ideal for someone who excels at technical troubleshooting, incident investigation, and customer communication in a B2B environment.</p>\n<p>As a key member of the support team, you will be responsible for handling escalated technical issues from on-premise enterprise clients, reproducing complex problems, and collaborating with engineering, data, and product teams to ensure swift resolution. You will report directly to the Head of Support, and play a critical role in maintaining customer satisfaction and improving our support operations.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Frontline Investigation: Handle escalated tickets from enterprise clients via Intercom, focusing on on-premise infrastructure and AI-related issues (e.g., deployment, performance, integration, security).</li>\n<li>Root Cause Analysis: Ask the right questions to gather context, reproduce issues in test environments, and diagnose technical problems (systems, networks, storage, GPU clusters, AI models).</li>\n<li>Cross-Team Collaboration: Work closely with engineering, and deployment teams to escalate, track, and resolve incidents efficiently.</li>\n<li>Proactive Communication: Provide clear, empathetic, and timely updates to clients and internal stakeholders, ensuring transparency throughout the resolution process.</li>\n</ul>\n<p>Knowledge Sharing &amp; Process Improvement:</p>\n<ul>\n<li>Documentation: Create and update technical FAQs, troubleshooting guides, and internal knowledge base articles to empower self-serve/L1 team and reduce recurrence of issues.</li>\n<li>Feedback Loop: Identify recurring pain points in on-premise deployments and suggest improvements to product, documentation, or support workflows.</li>\n</ul>\n<p>Customer-Centric Approach:</p>\n<ul>\n<li>Empathy &amp; Ownership: Maintain a customer-first mindset, ensuring clients feel heard and supported, even in high-pressure situations.</li>\n<li>Solution-Oriented: Proactively propose workarounds, fixes, or process optimizations to enhance the customer experience and reduce incident resolution time.</li>\n</ul>\n<p>Technical Expertise:</p>\n<ul>\n<li>On-Premise &amp; Cloud Environments: Deep understanding of Linux/Windows servers, networking, virtualization, storage, security (firewalls, RGPD compliance), and cloud providers (AWS, GCP, Azure).</li>\n<li>Kubernetes/Helm: Experience with deployment, scaling, and troubleshooting of applications in Kubernetes clusters using Helm charts.</li>\n<li>Terraform: Familiarity with Infrastructure as Code (IaC) for managing cloud resources is a strong plus.</li>\n<li>AI Infrastructure: Knowledge of AI/ML pipelines, LLM/RAG deployments, GPU acceleration, and data storage solutions for enterprise clients.</li>\n<li>Tooling: Proficiency in Intercom, monitoring tools, scripting (Bash/Python), and diagnostic utilities (logs, performance metrics).</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6f25b435-69f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Mistral AI","sameAs":"https://mistral.ai/careers","logo":"https://logos.yubhub.co/mistral.ai.png"},"x-apply-url":"https://jobs.lever.co/mistral/f00a13aa-61f1-4c56-993c-20846adc2b15","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Linux/Windows servers","Networking","Virtualization","Storage","Security","Kubernetes/Helm","Terraform","AI/ML pipelines","LLM/RAG deployments","GPU acceleration","Data storage solutions","Intercom","Monitoring tools","Scripting (Bash/Python)","Diagnostic utilities"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:47:50.345Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Paris"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Linux/Windows servers, Networking, Virtualization, Storage, Security, Kubernetes/Helm, Terraform, AI/ML pipelines, LLM/RAG deployments, GPU acceleration, Data storage solutions, Intercom, Monitoring tools, Scripting (Bash/Python), Diagnostic utilities"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_058f6a10-283"},"title":"Field Hardware Engineer, HPC","description":"<p>Our compute footprint is growing fast to support our science and engineering teams. We&#39;re hiring a Field HW Engineer to understand end-to-end systems, execute complex/vendor-level interventions, and guide L1 engineers on site,without direct line management.</p>\n<p>You&#39;ll work hands-on across compute, storage, interconnect and cooling to keep one of France&#39;s largest GPU/CPU clusters healthy and scalable.</p>\n<p>Location: Bruyères-le-Châtel , on-site, field role (multi-site mobility: Paris area and nearby)</p>\n<p>Reporting line: Hardware Ops</p>\n<p>Impact:</p>\n<p>• Compute is a key lever for Mistral&#39;s success and our largest spend item.</p>\n<p>• Direct impact on scale: you&#39;ll restore service on complex incidents and raise the bar on reliability as we grow.</p>\n<p>• Enable breakthrough AI: your work unlocks science &amp; engineering teams to deliver state-of-the-art AI.</p>\n<p>What you will do:</p>\n<p>• Lead complex interventions: plan and execute vendor-level or multi-node operations (e.g., full rack work, intricate recabling, post-restart diagnosis), own risk assessment/rollback, and coordinate with vendors (RMA/escalations).</p>\n<p>• Advanced diagnostics: correlate symptoms across compute, storage, interconnect, cooling; read system indicators (LED/POST/beep), BMC/IPMI consoles, and logs to identify root causes.</p>\n<p>• Guide and uplift L1s: coach on safe practices (ESD/LOTO), first-line triage, rack craftsmanship, documentation quality; pair on tricky procedures.</p>\n<p>• Process &amp; automation: improve SOPs/checklists; propose/build small automation (Python/Bash) for photo/serial capture, inventory sync, dashboards/alerts; shorten MTTR.</p>\n<p>• Safety &amp; compliance: enforce lockout/tagout, ESD, PPE; ensure audit-ready tickets, evidence and change traces.</p>\n<p>• Parts &amp; logistics (advanced): plan spares strategy, track failure trends, and drive proactive vendor actions.</p>\n<p>About you:</p>\n<p>• 5+ years in data center/server hardware or L2/L3 hardware support, with proven complex hands-on work in production (HPC/AI/Cloud at scale).</p>\n<p>• End-to-end hardware expertise: comfortable across CPU/memory/PCIe cards (incl. accelerators), NICs, PSUs, drives, network, power and cooling (including DLC); strong judgment on when/how to escalate.</p>\n<p>• Diagnostics depth: confident in analyzing BMC/IPMI logs, linux software logs and crashes simple CLI checks; methodical root cause analysis.</p>\n<p>• Safety &amp; discipline: impeccable ESD/LOTO/PPE habits; zero rough handling; clean, labeled, auditable work.</p>\n<p>• Communication &amp; mentoring: crisp status/handovers; able to coach L1s during live operations.</p>\n<p>Provide technical documentations to L1s or other team</p>\n<p>Mobility: willing to travel between sites (Paris area or nearby regions, occasionally in Europe or US))</p>\n<p>Nice to have:</p>\n<p>• Vendor tools (iDRAC/iLO/IPMI), RAID/storage basics (NVMe/SAS/SATA), high-speed interconnect (Ethernet/InfiniBand).</p>\n<p>• Coding/automation (Python/Bash) for small ops tools and reporting.</p>\n<p>• Experience with ticketing (Jira/ServiceNow), inventory/RMA flows, vendor coordination.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_058f6a10-283","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Mistral AI","sameAs":"https://mistral.ai","logo":"https://logos.yubhub.co/mistral.ai.png"},"x-apply-url":"https://jobs.lever.co/mistral/ea94b55b-58e1-437b-bf3d-07ed150308e3","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data center/server hardware","L2/L3 hardware support","HPC/AI/Cloud at scale","end-to-end hardware expertise","diagnostics depth","safety & discipline","communication & mentoring"],"x-skills-preferred":["vendor tools","RAID/storage basics","high-speed interconnect","coding/automation","ticketing","inventory/RMA flows","vendor coordination"],"datePosted":"2026-04-17T12:47:46.512Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Paris"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data center/server hardware, L2/L3 hardware support, HPC/AI/Cloud at scale, end-to-end hardware expertise, diagnostics depth, safety & discipline, communication & mentoring, vendor tools, RAID/storage basics, high-speed interconnect, coding/automation, ticketing, inventory/RMA flows, vendor coordination"}]}