{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/sql"},"x-facet":{"type":"skill","slug":"sql","display":"Sql","count":100},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_eec96030-9d4"},"title":"Senior Backend Engineer","description":"<p>Are you ready to revolutionize the way people experience banking?</p>\n<p>At bunq, we’re here to make life easier and redefine what banking can be. This is your chance to shape the future of digital banking by creating the best app experience imaginable. Use your coding prowess to build and enhance the Bank of The Free, ensuring seamless, secure, and flawless functionality for our users.</p>\n<p><strong>Take Ownership</strong></p>\n<p>As bunq&#39;s Senior Backend Engineer, you’ll be at the forefront of our backend development, ensuring our employees and users experience unparalleled stability and performance.</p>\n<ul>\n<li>Collaborate with our product team to design and implement robust backend structures for cutting-edge, user-facing features, utilizing technologies like MySQL, Redis, Elasticsearch, and PHP.</li>\n<li>Provide continuity and growth for our backend team by mentoring and guiding junior developers, helping them unlock their full potential.</li>\n<li>Manage high-availability production systems and support other departments by developing and maintaining efficient, scalable processes.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Our PHP-based backend is intuitive, so no prior PHP experience is needed while experience in any object-oriented language will do.</li>\n<li>You&#39;re able to build your own projects and manage high-availability production systems.</li>\n<li>You use your creativity to build industry breaking user facing features.</li>\n<li>You coordinate efforts to increase efficiency and scalability.</li>\n<li>You&#39;re fluent in English, both speaking and writing.</li>\n</ul>\n<p><strong>Your space to perform</strong></p>\n<p>We give you the space and the tools you need to succeed. Our team is great, international colleagues who share your mindset. We offer a hybrid setup: after 3 months in-office, work 2 days remote, 3 days in-office weekly. We also have a Digital Nomad Program: After your first year, enjoy up to 20 days per year to work while traveling, combining flexibility with strong team collaboration. We reward tenure with a dedicated travel budget: €1.5k after 2 years and €3k after 4 years to visit another core office.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_eec96030-9d4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"bunq","sameAs":"https://careers.bunq.com","logo":"https://logos.yubhub.co/careers.bunq.com.png"},"x-apply-url":"https://careers.bunq.com/o/senior-backend-engineer-4","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MySQL","Redis","Elasticsearch","PHP","object-oriented language"],"x-skills-preferred":[],"datePosted":"2026-04-19T13:29:53.671Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"MySQL, Redis, Elasticsearch, PHP, object-oriented language"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d7f00fa5-7a4"},"title":"Backend Engineer","description":"<p>Are you ready to revolutionize the way people experience banking?</p>\n<p>At bunq, we&#39;re here to make life easier and redefine what banking can be. Play a key role in delivering an exceptional banking app experience by ensuring seamless, secure, and flawless functionality for our users. Use your coding skills to build and improve bank of the Free from behind the scenes.</p>\n<p><strong>Take Ownership</strong></p>\n<p>As a Backend Engineer at bunq, you&#39;ll be at the forefront of our backend development, ensuring our employees and users experience unparalleled stability and performance.</p>\n<ul>\n<li>Building (user-facing) features.</li>\n<li>Collaborating with product to propose and designing the backend structure for new features and supporting other departments by developing and maintaining processes.</li>\n<li>You&#39;ll utilize technologies like MySQL, Redis, Elasticsearch, and PHP to develop the bank of The Free.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Our PHP-based backend is intuitive, so no prior PHP experience is needed while experience in any object-oriented language will do.</li>\n<li>You&#39;re able to tinker and test your own projects.</li>\n<li>You use your creativity to build industry-breaking user-facing features.</li>\n<li>You coordinate efforts to increase efficiency and scalability.</li>\n<li>You&#39;re fluent in English, both speaking and writing.</li>\n</ul>\n<p><strong>Your space to perform</strong></p>\n<p>We give you the space and the tools you need to succeed.</p>\n<p>Join forces with great colleagues across the globe to revolutionize banking.</p>\n<p>Make lasting impact by working on complex &amp; exciting challenges.</p>\n<p>Accelerate your career growth with bunq Academy and €1500 learning budget and additional study leave.</p>\n<p>Monthly contributions to your phone and internet bills.</p>\n<p>A Macbook to keep with you while you&#39;re with us.</p>\n<p>Hybrid setup: after 3 months in-office, work 2 days remote, 3 days in-office weekly.</p>\n<p>Digital Nomad Program: After your first year, enjoy up to 20 days per year to work while traveling, combining flexibility with strong team collaboration</p>\n<p>We reward tenure with a dedicated travel budget: €1.5k after 2 years and €3k after 4 years to visit another core office.</p>\n<p>Travel expenses are covered whether you come walking or by bike, bus or car (though we prefer green choices ).</p>\n<p>Fabulous in-house chefs cooking delicious lunches.</p>\n<p>An optional pension plan with monthly contribution from bunq (Netherlands).</p>\n<p>Private Health Insurance, just in case (Istanbul).</p>\n<p>A massive discount with Urban Sports Club for your wellbeing (Netherlands).</p>\n<p>Stay fit &amp; healthy with a Multisport gym card (Istanbul).</p>\n<p>Friday drinks, team events, and other celebrations - bunq style!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d7f00fa5-7a4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"bunq","sameAs":"https://careers.bunq.com","logo":"https://logos.yubhub.co/careers.bunq.com.png"},"x-apply-url":"https://careers.bunq.com/o/backend-engineer-5","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MySQL","Redis","Elasticsearch","PHP","object-oriented language"],"x-skills-preferred":[],"datePosted":"2026-04-19T13:29:45.273Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"MySQL, Redis, Elasticsearch, PHP, object-oriented language"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_61234903-9fa"},"title":"Engineering Manager (Java or Typescript) - Guest Experience (all genders)","description":"<p>Join our Guest Experience department as an Engineering Manager, leading a dynamic team focused on enhancing the search experience of our users.</p>\n<p>As an Engineering Manager, you will be part of the Discovery team in the Guest Experience department. The team is responsible for designing and maintaining the list page of our website, ensuring users can easily find the best vacation rental from our search results.</p>\n<p>Your contributions will help create a seamless and joyful journey for travellers, which will result in increasing conversion rates and customer satisfaction.</p>\n<p>Your team will consist of frontend &amp; backend engineers (direct reports), a project manager and a QA engineer.</p>\n<p>You&#39;ll work closely with the Ranking, Conqueror, and Marketing teams, which manage the machine learning models for property ranking on the list page, booking systems, and Holidu&#39;s marketing efforts. Together, you&#39;ll ensure a seamless and cohesive user experience.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Frontend: Typescript and NodeJS processes in Kubernetes. We use ReactJS, Zustand and TailwindCSS on the client and Express on the server.</li>\n</ul>\n<ul>\n<li>Backend: Java 17/21, Kotlin (Spring Boot).</li>\n</ul>\n<ul>\n<li>Infrastructure: Microservices architecture deployed on AWS Kubernetes (EKS).</li>\n</ul>\n<ul>\n<li>Data Management: PostgreSQL, Redis, Elasticsearch 7, Redshift (part of a data lake structure).</li>\n</ul>\n<ul>\n<li>DevOps Tools: AWS, Docker, Jenkins, Git, Terraform.</li>\n</ul>\n<ul>\n<li>Monitoring &amp; Analytics: ELK, Grafana, Looker, Opsgenie, and in-house solutions.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<ul>\n<li>Lead a high-performing cross-functional team, focusing on product innovation, infrastructure reliability, delivery speed, quality, engineering culture, and team growth.</li>\n</ul>\n<ul>\n<li>Ensure your team delivers applications that are highly scalable, highly available, and capable of handling high traffic of up to 1 million unique users per day.</li>\n</ul>\n<ul>\n<li>Support team growth through regular feedback, mentorship, and by recruiting exceptional engineers.</li>\n</ul>\n<ul>\n<li>Work closely with product management, product design, and stakeholders to define the team&#39;s goals (OKR’s) and roadmap.</li>\n</ul>\n<ul>\n<li>Collaborate with peers, staff engineers, and other stakeholders to drive strategic technology decisions.</li>\n</ul>\n<ul>\n<li>Lead strategic team-driven projects, identify opportunities, define and uphold quality standards.</li>\n</ul>\n<ul>\n<li>Foster a great team culture aligned with the company values, ownership, autonomy, and inclusivity within your team and the entire department.</li>\n</ul>\n<ul>\n<li>Take full responsibility for delivering impactful features to millions of users annually.</li>\n</ul>\n<p>The role includes dedicating approximately 40-50% of the time as an individual contributor focused on feature implementation.</p>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>A bachelor&#39;s degree in Computer Science, a related technical field or equivalent practical experience.</li>\n</ul>\n<ul>\n<li>Experience building and implementing backend services and/or frontend applications.</li>\n</ul>\n<ul>\n<li>Experience providing technical leadership (e.g., setting goals and priorities, architecture design, task planning and code reviews).</li>\n</ul>\n<ul>\n<li>Experience as a people manager with the ability to build an excellent team culture based on mutual respect, empathy, learning and support for each other.</li>\n</ul>\n<ul>\n<li>Love for building world-class products with a great user experience.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters,and you’ll see the impact.</li>\n</ul>\n<ul>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets,with a strong focus on AI.</li>\n</ul>\n<ul>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts,people we can all relate to,making work meaningful and energizing.</li>\n</ul>\n<ul>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n</ul>\n<ul>\n<li>Flexibility:  Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n</ul>\n<ul>\n<li>Competitive Package: 95.000-125.000€ + VSOPs based on relevant experience and seniority , learn more about our approach to compensation here.</li>\n</ul>\n<ul>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized,but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_61234903-9fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/1558189","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":"95.000-125.000€ + VSOPs based on relevant experience and seniority","x-skills-required":["Typescript","NodeJS","ReactJS","Zustand","TailwindCSS","Express","Java","Kotlin","Spring Boot","AWS","Docker","Jenkins","Git","Terraform","PostgreSQL","Redis","Elasticsearch","Redshift","ELK","Grafana","Looker","Opsgenie"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:57.912Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Typescript, NodeJS, ReactJS, Zustand, TailwindCSS, Express, Java, Kotlin, Spring Boot, AWS, Docker, Jenkins, Git, Terraform, PostgreSQL, Redis, Elasticsearch, Redshift, ELK, Grafana, Looker, Opsgenie"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6e0788ce-490"},"title":"AI Product Manager","description":"<p>Job Title: AI Product Manager</p>\n<p>We are seeking an experienced AI Product Manager to lead AI-focused product enablement and adoption initiatives across our global organisation. The ideal candidate will have a mix of technical experience in AI/ML/LLM products and APIs, process automation, and proficiency in survey design and data analytics.</p>\n<p>Key Responsibilities:</p>\n<p>Awareness of AI Ecosystem and Trends: Use Claude Code / Codex / Cursor / etc. as well as skills in these tools. Follow developments among major players in the commercial and open-source AI/LLM space, including Anthropic, OpenAI, Google, xAI, and others.</p>\n<p>Communication: Communicate clearly and effectively with internal and external vendors and stakeholders about requirements, feature requests, scope, expectations, priorities, product releases, and timelines. Generate clear, consistent, and accurate documentation about products, tradeoffs, decisions, and value proposition of various efforts.</p>\n<p>AI Product Enablement and Adoption: Conduct consulting-style engagements with technical and non-technical teams to onboard them with AI tools and products available within Millennium. Collaborate with users to understand pain points, needs, feature requests, and requirements to design and execute product development with usability and scalability in mind.</p>\n<p>Feedback Collection and Data Analytics: Analyze usage and feedback data with accuracy and quality using SQL and Python to identify trends, gaps, and opportunities for product improvements. Use insights from your analysis to refine product roadmaps and enablement initiatives to maximise impact across the firm.</p>\n<p>Qualifications:</p>\n<p>Education: Bachelor&#39;s degree or higher in Computer Science, Data Science, Engineering, or a related technical field. Must have technical knowledge and experience.</p>\n<p>Experience: 5-7+ years of experience in a combination of AI/ML/LLM engineering, startup, technical education, consulting, data science, product analytics, or product management roles. Some technical / engineering / building experience is a hard requirement, e.g. you must be able to use the CLI and analyse data.</p>\n<p>Technical Proficiency: Strong knowledge of AI/ML concepts, especially Large Language Models (LLMs). Understanding of their capabilities and failings as a technology. Proficiency in SQL and Python for data analysis and visualisation. Proficiency in survey software and analytics tools (e.g. Qualtrics). Strong familiarity with effective SDLC and CI/CD principles.</p>\n<p>Product Management: Proven ability to use clear judgment and organisational skills to manage complex, cross-functional products involving technical and non-technical teams. Intellectual curiosity and strong aptitude for prioritisation.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6e0788ce-490","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT Infrastructure","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955422287","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Claude Code","Codex","Cursor","SQL","Python","Qualtrics","SDLC","CI/CD"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:37.294Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore, Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Technology","skills":"Claude Code, Codex, Cursor, SQL, Python, Qualtrics, SDLC, CI/CD"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b33cbd91-bc9"},"title":"Systematic Production Support Engineer","description":"<p>We are seeking an experienced Systematic Production Support Engineer to help us scale our systematic operations and support engineering capabilities. This role directly supports portfolio management teams across Millennium, with operational excellence at the core. Our efforts are focused on delivering the highest quality returns to our investors – providing a world-class and reliable trading and technology platform is essential to this mission.</p>\n<p>As a Systematic Production Support Engineer, you will be responsible for building, developing, and maintaining a reliable, scalable, and integrated platform for trading strategy monitoring, reporting, and operations. You will work closely with portfolio managers and other internal customers to reduce operational risk through the implementation of monitoring, reporting, and trade workflow solutions, as well as automated systems and processes focused on trading and operations.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Building, developing, and maintaining a reliable, scalable, and integrated platform for trading strategy monitoring, reporting, and operations</li>\n<li>Working with portfolio managers and other internal customers to reduce operational risk through the implementation of monitoring, reporting, and trade workflow solutions</li>\n<li>Implementing automated systems and processes focused on trading and operations</li>\n<li>Streamlining development and deployment processes</li>\n</ul>\n<p>Technical qualifications include:</p>\n<ul>\n<li>5+ years of development experience in Python</li>\n<li>Experience working in a Linux/Unix environment</li>\n<li>Experience working with PostgreSQL or other relational databases</li>\n</ul>\n<p>Preferred skills and experience include:</p>\n<ul>\n<li>Understanding of NLP, supervised/non-supervised learning, and Generative AI models</li>\n<li>Experience operating and monitoring low-latency trading environments</li>\n<li>Familiarity with quantitative finance and electronic trading concepts</li>\n<li>Familiarity with financial data</li>\n<li>Broad understanding of equities, futures, FX, or other financial instruments</li>\n<li>Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#</li>\n<li>Experience with Apache/Confluent Kafka</li>\n<li>Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline)</li>\n<li>Experience with containerization and orchestration technologies</li>\n<li>Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure</li>\n<li>Contributions to open-source projects</li>\n</ul>\n<p>This is a unique opportunity to drive significant value creation for one of the world&#39;s leading investment managers.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b33cbd91-bc9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Unknown","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954716155","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Linux/Unix","PostgreSQL","NLP","supervised/non-supervised learning","Generative AI models","low-latency trading environments","quantitative finance","electronic trading concepts","financial data","equities","futures","FX","distributed systems","backend development","C/C++","Java","Scala","Go","C#","Apache/Confluent Kafka","SDLC pipelines","containerization","orchestration technologies","AWS","GCP","Azure"],"x-skills-preferred":["Understanding of NLP, supervised/non-supervised learning, and Generative AI models","Experience operating and monitoring low-latency trading environments","Familiarity with quantitative finance and electronic trading concepts","Familiarity with financial data","Broad understanding of equities, futures, FX, or other financial instruments","Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#","Experience with Apache/Confluent Kafka","Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline)","Experience with containerization and orchestration technologies","Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure","Contributions to open-source projects"],"datePosted":"2026-04-18T22:14:36.583Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Linux/Unix, PostgreSQL, NLP, supervised/non-supervised learning, Generative AI models, low-latency trading environments, quantitative finance, electronic trading concepts, financial data, equities, futures, FX, distributed systems, backend development, C/C++, Java, Scala, Go, C#, Apache/Confluent Kafka, SDLC pipelines, containerization, orchestration technologies, AWS, GCP, Azure, Understanding of NLP, supervised/non-supervised learning, and Generative AI models, Experience operating and monitoring low-latency trading environments, Familiarity with quantitative finance and electronic trading concepts, Familiarity with financial data, Broad understanding of equities, futures, FX, or other financial instruments, Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#, Experience with Apache/Confluent Kafka, Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline), Experience with containerization and orchestration technologies, Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure, Contributions to open-source projects"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b93800dd-3d2"},"title":"Production Engineering Support Manager – Liquidity Provision Technology","description":"<p>We are seeking a Production Engineering Support Manager to join our team. As a Production Engineering Support Manager, you will provide leadership and guidance to coach, motivate and lead team members to their optimum performance levels and career development. You will solve technical trading-related issues, independently where possible or leveraging teammates as necessary. You will escalate to application and/or infrastructure subject matter experts (internally or at vendors) when appropriate. You will manage communications to the trading staff and internal stakeholders, primarily our execution services team regarding issue/resolution.</p>\n<p>Collaborate with other technical support engineers who may need assistance working on an issue; utilize his/her area of expertise to help others in order to quickly facilitate solutions for the customer. Build and foster working relationships with trading groups with a focus on execution services team. Work with global counterparts to provide seamless 24/7 global coverage.</p>\n<p>Trading Infrastructure / Platform Status Communications – disseminate messages to the appropriate trading staff regarding trading infrastructure / platform issues, exchange updates, etc. Uplift environment management tools to reduce risk and streamline efficiency of support team. Assist with automating processes to achieve efficiency and streamlined trade support.</p>\n<p>Document and create new knowledge base to provide the most effective solutions to trading issues. Deployment of, support of, and monitoring of the firm’s internal trading systems. Coordinate with vendors, internal application owners, infrastructure owners, and tech support to ensure trading platforms are correctly installed, configured, and tested.</p>\n<p>Liaise with development and infrastructure teams, prioritize tool enhancements, and coordinate and participate in software/new version releases. Liaise with Dev and Infrastructure teams to coordinate and participate in software/new version releases.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b93800dd-3d2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755953129734","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Linux","shell scripting","python","SQL","financial technology","FIX protocols","AI technologies","version control systems","SDLC processes","columnar database","AWS"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:31.477Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Linux, shell scripting, python, SQL, financial technology, FIX protocols, AI technologies, version control systems, SDLC processes, columnar database, AWS","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d676e040-d22"},"title":"Operations Specialist (Corporate Actions)","description":"<p>Operations Specialist (Corporate Actions)</p>\n<p>The Operations SpecialistMBED will be responsible for monitoring, coordinating, booking and communicating all Corporate Actions to the relevant trading teams. This role will also involve daily responsibilities revolving around desk support, same day trade matching, full reconciliations and PM/broker/counterparty queries as well as ad-hoc PnL issues relating to corporate action bookings and entitlements.</p>\n<p>Principal Responsibilities</p>\n<p>Actively managing the Corporate Action process for the EMEA markets across all event types Interacting with trading groups across multiple strategies accommodating their specific requirements, as well as several Prime Brokers (PBs) and potential vendors Processing cross-border deals and understanding mechanics on such deals including funding queries, FX risk and PnL exposure Able to actively monitor and process paydate process including bookings made to internal lines, pricing appropriately and working with PnL / valuation teams Intraday trade monitoring - working with PMs/Execution desk/IT to ensure correct capture Mitigating risk by ensuring timely completion of daily trade / cash / position / market value reconciliations Working closely with operations team based in India and acting as a focal point for Corporate Action related queries Educate and assist team members on all asset servicing issues Act as a primary contact for all EMEA related asset servicing queries and issues Assist the development of proprietary systems Continuous improvement of current systems working alongside regional teams and business analysts / technology</p>\n<p>Qualifications/Skills Required</p>\n<p>5+ years relevant experience in Asset Servicing/Corporate Action team and/or equity operations/ support Strong knowledge of Corporate Actions processing from front to back Must have a practical working knowledge of equity swap mechanics in a Buy vs Sell side environment Excellent excel skills and familiarity with SQL / Python Ability to implement and maintain new systems, procedures, and controls Strong and confident communication and interpersonal skills with clear ability to be able to face off to the trading desk and PMs Must be able to work under pressure and meet strict deadlines Omgeo / CTM / CTC / Traiana experience desired Detail oriented; Demonstrates thoroughness and strong ownership of work Must be able to work independently using well-honed analytical skills and abstract reasoning Good team player with a strong willingness to participate and help others Able to prioritize in a fast moving, high pressure, constantly changing environment; Good sense of urgency Willingness to undertake new challenges and opportunities should they present themselves</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d676e040-d22","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Core Operations","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955300494","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Corporate Actions processing","Equity swap mechanics","Excel","SQL","Python","Omgeo","CTM","CTC","Traiana"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:27.064Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"Operations","industry":"Finance","skills":"Corporate Actions processing, Equity swap mechanics, Excel, SQL, Python, Omgeo, CTM, CTC, Traiana"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5c943be1-a6f"},"title":"Threat Detection Engineer","description":"<p>We are seeking an experienced Threat Detection Engineer to join our team. As a Threat Detection Engineer, you will be responsible for driving our best-in-class posture through threat hunting, detection engineering, and playbook creation. You will also lead information security response activities and work closely with business and technology teams to deliver positive outcomes.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Leveraging multiple data sources to identify modern evolving threats and developing new detection and response approaches</li>\n<li>Creating and operating high-fidelity detections mechanisms that drive efficient, effective, and repeatable response</li>\n<li>Owning, operating, and automating detection and response workflows to enable the team to focus on strategic objectives</li>\n<li>Leading information security response activities for the firm</li>\n<li>Working across business and technology teams to deliver positive outcomes across the firm</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>A bachelor&#39;s or master&#39;s degree in computer science or cyber security with a strong IT background or equivalent demonstrable experience</li>\n<li>3 years&#39; experience working in a security engineering role, with financial industry experience preferred</li>\n<li>Experience in creating detections in modern query languages (KQL, SQL, SPL)</li>\n<li>Possession of security certifications (Security+, OSCP, CISSP, CEH, GCIA, GCIH)</li>\n<li>Experience with modern security tooling across security domains; network, endpoint, data, identity, and cloud</li>\n<li>Experience in standard enterprise technology stack, including Active Directory, Entra, Group Policy, Intune, DNS, TCP/IP, PKI, Microsoft 365, Windows, Linux, MacOS, etc.</li>\n<li>Ability to handle sensitive and/or confidential materials with appropriate discretion</li>\n<li>Required scripting, development, and automation skills using PowerShell or Python and proficient development tools</li>\n<li>Experience in OSINT, threat hunting, and analysing malicious emails</li>\n<li>Ability to prioritise in a fast-moving, high-pressure, constantly changing environment</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5c943be1-a6f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT Infrastructure","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955470563","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["threat detection","security engineering","cyber security","KQL","SQL","SPL","security certifications","modern security tooling","standard enterprise technology stack","scripting","development","automation","OSINT","threat hunting"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:24.700Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Finance","skills":"threat detection, security engineering, cyber security, KQL, SQL, SPL, security certifications, modern security tooling, standard enterprise technology stack, scripting, development, automation, OSINT, threat hunting"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_86156963-6d9"},"title":"Commodities Finance Manager","description":"<p>We are seeking a highly skilled Commodities Finance Manager to join our Fund Accounting team. The ideal candidate will possess deep expertise in commodities markets, fund accounting principles, and operational processes within a hedge fund environment.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Oversee the accounting and reconciliation of commodities-related transactions, including futures, options, swaps, and physical commodities.</li>\n<li>Ensure timely and accurate preparation of monthly, and quarterly NAV calculations for commodities entities.</li>\n<li>Collaborate with the broader accounting team to ensure compliance with internal policies and external regulatory requirements.</li>\n<li>Work closely with trading desks, operations, and counterparties to resolve discrepancies and ensure accurate reporting.</li>\n<li>Provide insights into market trends and their impact on portfolio valuation and risk metrics.</li>\n<li>Identify and implement process improvements to enhance the efficiency and accuracy of commodities accounting workflows.</li>\n<li>Leverage technology and automation tools to streamline reporting and reconciliation processes.</li>\n<li>Ensure adherence to relevant regulatory frameworks, including GAAP, IFRS, and other applicable standards.</li>\n<li>Prepare audit documentation and liaise with external auditors as needed.</li>\n</ul>\n<p>Qualifications/Skills Required:</p>\n<ul>\n<li>Education: Bachelor’s degree in Accounting, Finance, Economics, or a related field.</li>\n<li>Experience: 10+ years of experience in fund accounting, with a focus on commodities markets.</li>\n<li>Technical Skills: Proficiency in fund accounting systems (e.g., Geneva).</li>\n<li>Advanced Excel skills; familiarity with data visualization tools and programming languages (e.g., Python, SQL) is a plus.</li>\n<li>Strong understanding of commodities markets, including derivatives and physical assets.</li>\n<li>Familiarity with regulatory requirements impacting hedge funds and commodities trading.</li>\n</ul>\n<p>Why Join Millennium Management?</p>\n<p>Millennium Management is a premier global hedge fund with a reputation for excellence and innovation. As a Commodities Manager in Fund Accounting, you will have the opportunity to work alongside some of the brightest minds in the industry, contribute to the success of a world-class investment platform, and advance your career in a dynamic and rewarding environment.</p>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package. The estimated base salary range for this position is $160,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_86156963-6d9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Fund Accounting","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755953514852","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 to $250,000","x-skills-required":["fund accounting","commodities markets","GAAP","IFRS","Geneva","Excel","Python","SQL"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:21.927Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"fund accounting, commodities markets, GAAP, IFRS, Geneva, Excel, Python, SQL","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d96db681-39c"},"title":"Operations Control Specialist","description":"<p>We are building a specialized team within Operations focused on designing and implementing controls and analytics for trade surveillance and trade processing.</p>\n<p>The Operations Control Specialist will combine technical skills and business knowledge to build control metrics, dashboards, and automated checks on surveillance and trade data.</p>\n<p>This role offers exposure to products across multiple asset classes, trading platforms, exchange symbology, and global markets.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Design, implement, and automate reconciliations and health checks across large trade and surveillance datasets (e.g., OMS and upstream/downstream systems).</li>\n</ul>\n<ul>\n<li>Build monitoring dashboards and control indicators (metrics/KPIs) to validate that trades are correctly generated, routed, and reported.</li>\n</ul>\n<ul>\n<li>Develop and maintain data quality checks (completeness, accuracy, timeliness) and automated exception reporting for trade and alert data.</li>\n</ul>\n<ul>\n<li>Partner with Technology on system migrations and application redesign, defining control requirements and validating outcomes through testing and analytics.</li>\n</ul>\n<ul>\n<li>Support data governance efforts by identifying data issues, documenting data lineage, and contributing to standards for critical trade and surveillance data.</li>\n</ul>\n<ul>\n<li>Design and automate reporting using data analytics, automation, and AI-enabled tools to reduce manual processes and improve transparency.</li>\n</ul>\n<ul>\n<li>Collaborate with Operations, Compliance, and Middle Office stakeholders to understand workflows, refine control logic, and address issues identified by analytics.</li>\n</ul>\n<p>Qualifications &amp; Skills:</p>\n<ul>\n<li>5+ years of experience in the financial industry, ideally in Operations, Middle Office, Risk, or Surveillance/Compliance analytics.</li>\n</ul>\n<ul>\n<li>Strong SQL skills (e.g., SQL, PL/SQL, T‑SQL) and experience working with large, complex datasets.</li>\n</ul>\n<ul>\n<li>Programming experience in Python (or a similar language) for data analysis, automation, and scripting.</li>\n</ul>\n<ul>\n<li>Familiarity with institutional trading workflows and trading / surveillance technology (e.g., order management systems, trade reporting, or surveillance platforms).</li>\n</ul>\n<ul>\n<li>Experience with data visualization, reporting, and analytics tools used to build dashboards and control reporting.</li>\n</ul>\n<ul>\n<li>Strong analytical and problem-solving skills; able to diagnose data and process issues and propose practical solutions.</li>\n</ul>\n<ul>\n<li>Excellent communication skills, with the ability to explain complex data and technical concepts to non-technical stakeholders.</li>\n</ul>\n<ul>\n<li>Strong interpersonal skills and comfort working with cross-functional teams (Operations, Compliance, Technology, Front Office).</li>\n</ul>\n<ul>\n<li>Highly organized self-starter with the ability to prioritize, manage multiple tasks, and take end-to-end ownership.</li>\n</ul>\n<ul>\n<li>Detail-oriented and proactive in identifying, investigating, and resolving data or process issues.</li>\n</ul>\n<ul>\n<li>Basic to intermediate understanding of financial instruments and products across asset classes.</li>\n</ul>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p>The estimated base salary range for this position is $160,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d96db681-39c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Regulatory Reporting Ops","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955532760","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$160,000 to $250,000","x-skills-required":["SQL","Python","Data analysis","Automation","Scripting","Institutional trading workflows","Trading / surveillance technology","Data visualization","Reporting","Analytics","Analytical skills","Problem-solving skills","Communication skills","Interpersonal skills"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:21.906Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"SQL, Python, Data analysis, Automation, Scripting, Institutional trading workflows, Trading / surveillance technology, Data visualization, Reporting, Analytics, Analytical skills, Problem-solving skills, Communication skills, Interpersonal skills","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6722b21a-d69"},"title":"AI Product Manager - Infrastructure","description":"<p>The AI Product Manager role within Infrastructure at Millennium will lead AI-focused product enablement and adoption initiatives across our global organization.</p>\n<p>The ideal candidate will have a mix of technical experience in AI/ML/LLM products and APIs, process automation, and proficiency in survey design and data analytics.</p>\n<p>They will play a key role in identifying pain points in the organization, understanding the landscape of existing AI solutions in the market (or what can be built internally), matching solutions to those problems, and onboarding teams with new and existing AI tools to foster a culture of AI awareness and innovation across Millennium.</p>\n<p>Key Responsibilities:</p>\n<p>Awareness of AI Ecosystem and Trends - Use Claude Code / Codex / Cursor / etc. as well as Skills in these tools.</p>\n<p>Follow developments among major players in the commercial and open-source AI/LLM space, including Anthropic, OpenAI, Google, xAI, and others.</p>\n<p>Communication - Communicate clearly and effectively with internal and external vendors and stakeholders about requirements, feature requests, scope, expectations, priorities, product releases, and timelines.</p>\n<p>Generate clear, consistent, and accurate documentation about products, tradeoffs, decisions, and value proposition of various efforts.</p>\n<p>AI Product Enablement and Adoption - Conduct consulting-style engagements with technical and non-technical teams to onboard them with AI tools and products available within Millennium.</p>\n<p>Collaborate with users to understand pain points, needs, feature requests, and requirements to design and execute product development with usability and scalability in mind.</p>\n<p>Feedback Collection and Data Analytics - Analyze usage and feedback data with accuracy and quality using SQL and Python to identify trends, gaps, and opportunities for product improvements.</p>\n<p>Use insights from your analysis to refine product roadmaps and enablement initiatives to maximize impact across the firm.</p>\n<p>Required Skills/Qualifications:</p>\n<p>Bachelor’s degree or higher in Computer Science, Data Science, Engineering, or a related technical field.</p>\n<p>Must have technical knowledge and experience.</p>\n<p>5-7+ years of experience in a combination of AI/ML/LLM engineering, startup, technical education, consulting, data science, product analytics, or product management roles.</p>\n<p>Some technical / engineering / building experience is a hard requirement, e.g. you must be able to use the CLI and analyze data.</p>\n<p>Strong knowledge of AI/ML concepts, especially Large Language Models (LLMs).</p>\n<p>Understanding of their capabilities and failings as a technology.</p>\n<p>Proficiency in SQL and Python for data analysis and visualization.</p>\n<p>Proficiency in survey software and analytics tools (e.g. Qualtrics).</p>\n<p>Strong familiarity with effective SDLC and CI/CD principles.</p>\n<p>Proven ability to use clear judgment and organizational skills to manage complex, cross-functional products involving technical and non-technical teams.</p>\n<p>Intellectual curiosity and strong aptitude for prioritization.</p>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p>When finalizing an offer, we take into consideration an individual’s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6722b21a-d69","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT Infrastructure","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954012033","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["AI/ML/LLM products and APIs","Process automation","Survey design and data analytics","SQL and Python for data analysis and visualization","Survey software and analytics tools (e.g. Qualtrics)"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:18.605Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Technology","skills":"AI/ML/LLM products and APIs, Process automation, Survey design and data analytics, SQL and Python for data analysis and visualization, Survey software and analytics tools (e.g. Qualtrics)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_78270c8d-016"},"title":"Operations Data Governance & Controls Specialist","description":"<p>As an Operations Control Specialist – Data Governance &amp; Controls, you will design, implement, and support technical data governance solutions with a focus on the firm&#39;s Trader Master and related reference data domains.</p>\n<p>This role requires a strong technical background in Data Management, Data Architecture, Data Lineage, Data Quality, Master Data Management (MDM), and automation within Financial Services and/or Technology.</p>\n<p>You will contribute to and help lead the technical design of data governance controls, data models, and integration patterns, partnering closely with Technology and Operations teams.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Build/enhance data governance frameworks, controls, standards, and workflows (policies, definitions, entitlements).</li>\n<li>Create data quality rules and monitoring; automate exception detection, alerting, remediation, SLAs, and RCA.</li>\n<li>Develop Python/SQL/ETL-ELT automation for checks, controls, and reporting; deliver Tableau/Power BI dashboards and KPIs.</li>\n<li>Contribute to conceptual/logical/physical data modeling for Trader Master and core domains.</li>\n<li>Support MDM capabilities: golden record, matching/merging, survivorship, stewardship workflows; help shape MDM strategy.</li>\n<li>Implement access/entitlement governance (RBAC, row/column security) across DB/warehouse/BI with audit compliance.</li>\n<li>Maintain catalog, glossary, lineage, schema history, impact analysis; manage structured change workflows.</li>\n<li>Define integration patterns (batch/API/streaming) and build reconciliations/validations across systems.</li>\n<li>Manage historical/temporal data (validation, backfills, remediation) supporting regulatory/reporting/analytics.</li>\n<li>Produce technical documentation (designs, runbooks, data dictionaries), share knowledge, and mentor juniors.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Engineering, Information Systems, Mathematics, Finance, or related field; advanced degree (MS, MBA, or equivalent) is a plus.</li>\n<li>5–8 years of experience in financial services or fintech with hands-on work in data engineering, data management, or data architecture roles; exposure to trading strategies, fund structures, and financial products strongly preferred.</li>\n</ul>\n<p>Technical Expertise (Required):</p>\n<ul>\n<li>Strong Python and SQL; experience with data warehousing + ETL/ELT.</li>\n<li>Familiarity with MDM/data governance tools (e.g., Collibra, Informatica, Alation) and Tableau/Power BI.</li>\n<li>Proven ability to lead delivery, solve complex data issues, and communicate with technical/non-technical stakeholders.</li>\n<li>Preferred certs: DAMA/CDMP, cloud (AWS/Azure/GCP), Scrum, BI/data engineering.</li>\n</ul>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p>The estimated base salary range for this position is $70,000 to $160,000, which is specific to New York and may change in the future.</p>\n<p>When finalizing an offer, we take into consideration an individual’s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_78270c8d-016","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Ops & MO Control","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954926796","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$70,000 to $160,000","x-skills-required":["Python","SQL","ETL/ELT","Data Warehousing","Tableau/Power BI","MDM/data governance tools","Collibra","Informatica","Alation"],"x-skills-preferred":["DAMA/CDMP","cloud (AWS/Azure/GCP)","Scrum","BI/data engineering"],"datePosted":"2026-04-18T22:14:17.909Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, SQL, ETL/ELT, Data Warehousing, Tableau/Power BI, MDM/data governance tools, Collibra, Informatica, Alation, DAMA/CDMP, cloud (AWS/Azure/GCP), Scrum, BI/data engineering","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":160000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_59421d7b-b28"},"title":"Full Stack Engineer - Real-Time Trading","description":"<p>We are seeking a Full Stack Engineer to join our EQ Real-Time P&amp;L &amp; Risk team. This team is responsible for designing, developing, and supporting technology platforms that enable our businesses to view, evaluate, hedge, and trade live positions, P&amp;L, and risk.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Collaborate with application development teams, technology management, and the business to design, prototype, and implement next-generation web UIs and mobile apps.</li>\n<li>Develop, maintain, and support existing Java Client UI used by a quarter of the firm.</li>\n<li>Contribute to the application development and architecture of highly scalable real-time UIs.</li>\n</ul>\n<p>Qualifications, Skills, and Requirements:</p>\n<ul>\n<li>5+ years of full-stack development experience, preferably within a financial services firm supporting real-time UIs.</li>\n<li>Expertise with Core Java and Spring.</li>\n<li>Excellent grasp of data structures and algorithms and the ability to learn and adopt new technologies quickly.</li>\n<li>Familiarity with database technologies – Advanced SQL, NoSQL, Time-series databases (KDB).</li>\n<li>Experience with event-driven architecture using message bus and caching technologies like Solace, Kafka, Pulsar, Memcached, Redis.</li>\n<li>Experience working with various monitoring tools like Datadog, ELK stack.</li>\n<li>A strong interest in financial markets and a desire to work directly with investment professionals.</li>\n<li>A good team player with a strong willingness to participate and help others.</li>\n<li>Drive to learn and experiment.</li>\n</ul>\n<p>Nice-to-have:</p>\n<ul>\n<li>Proficiency with Angular UI is preferred; React will also be considered.</li>\n<li>Familiarity with equities and equity derivatives within a real-time electronic trading environment is preferred.</li>\n<li>Experience with KDB+ q or C/C++.</li>\n</ul>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_59421d7b-b28","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954774219","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Core Java","Spring","Advanced SQL","NoSQL","Time-series databases (KDB)","Solace","Kafka","Pulsar","Memcached","Redis","Datadog","ELK stack"],"x-skills-preferred":["Angular UI","React","equities","equity derivatives","KDB+ q","C/C++"],"datePosted":"2026-04-18T22:14:10.158Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Core Java, Spring, Advanced SQL, NoSQL, Time-series databases (KDB), Solace, Kafka, Pulsar, Memcached, Redis, Datadog, ELK stack, Angular UI, React, equities, equity derivatives, KDB+ q, C/C++","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7a90d311-fba"},"title":"Full Stack Engineer - Equities Autocallables","description":"<p>This role is part of a global team responsible for enhancing and supporting a real-time trade capture platform that processes, normalizes, and enriches the firm&#39;s executions across multiple asset classes. The platform feeds executions into downstream systems including real-time P&amp;L, risk, and reporting.</p>\n<p>The successful candidate will focus on a Private Credit buildout, with particular emphasis on equities and options, and on integrating with third-party platforms such as Murex and ION. They will design, develop, and maintain Java-based services that support a real-time trade capture platform for our autocallable buildout, and build and support Kafka-based streaming pipelines to process, normalize, and distribute trading and reference data to downstream systems.</p>\n<p>Key responsibilities include collaborating closely with portfolio managers, traders, operations, and risk teams to understand requirements and translate them into robust technical solutions, contributing to the architecture and design of low-latency, high-availability components, including multithreaded and distributed systems, and monitoring, troubleshooting, and resolving production issues related to trading workflows, data integrity, and system performance.</p>\n<p>We are looking for a highly skilled and experienced software engineer with a strong background in Java, Kafka, and front-end technologies using Typescript/Javascript, in this role you&#39;ll be using Angular. You should have a solid understanding of object-oriented design, design patterns, and multithreading in distributed systems, and hands-on experience with unit testing and integration testing frameworks and best practices.</p>\n<p>In addition, you should be familiar with CI/CD pipeline (Jenkins) and DevOps tools/practices (e.g., Git, build tools, automated testing, deployment automation), experience with SQL databases such as Postgres and SQLServer, and comfort with modern IDEs and developer productivity tools; openness to using AI-assisted development tools and modern developer workflows.</p>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7a90d311-fba","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954367614","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Java","Kafka","Angular","Typescript","Postgres","SQLServer","Jenkins","Git","CI/CD pipeline","DevOps"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:08.489Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America · New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Kafka, Angular, Typescript, Postgres, SQLServer, Jenkins, Git, CI/CD pipeline, DevOps","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_07c95966-8e7"},"title":"Backend Developer - Host Experience (all genders)","description":"<p>Join our Host Experience department as a Backend Developer and become part of the team that brings new vacation rental properties to life on Holidu.</p>\n<p>You&#39;ll be working at the heart of our property acquisition engine , where we take hosts from their very first sign-up all the way to their first booking, making that journey as fast and seamless as possible.</p>\n<p>This team sits at a uniquely strategic intersection of product and growth. You will build and optimize the systems that every new host flows through: from onboarding and listing creation, to property configuration, content quality, and referral programs.</p>\n<p>The work demands reliability and attention to detail , because the time between a host signing up and welcoming their first guest, and how well their property performs from day one, is directly shaped by the quality of what you build.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Backend written in Kotlin and Java 21+ (with Spring Boot), with Gradle.</li>\n<li>Deployed as microservices on AWS-hosted Kubernetes cluster (EKS).</li>\n<li>Internal and external web applications written with ReactJS.</li>\n<li>Event-driven communication between services through EventBridge with SQS / ActiveMQ.</li>\n<li>Usage of a diverse set of technologies depending on the use case, such as PostgreSQL, S3, Valkey, ElasticSearch, GraphQL, and many more.</li>\n<li>Monitoring with OpenTelemetry, Grafana, Prometheus, ELK, APM, and CloudWatch.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<ul>\n<li>Design, build, evolve, and maintain our services, creating a great user experience for our hosts.</li>\n<li>Build a strong understanding of the product, use it to drive initiatives end-to-end, and contribute to shaping the team&#39;s direction as you grow.</li>\n<li>Work AI-first: use AI to accelerate not just coding, but data exploration, codebase understanding, technical design, and decision-making , and continuously sharpen how you use these tools.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>A passion for great user experience and drive to deliver world-class products.</li>\n<li>Early experience delivering product impact through engineering , you&#39;ve shipped things that real users depend on.</li>\n<li>Experience with Java or Kotlin with Spring is a plus.</li>\n<li>Experience with relational databases and deploying apps in cloud environments. NoSQL experience is a plus.</li>\n<li>Familiarity with various API types and integration best practices.</li>\n<li>Strong problem-solving skills and a team-oriented mindset.</li>\n<li>Curiosity for the business side - you want to understand the “why” behind the features.</li>\n<li>A love for coding and building high-quality products that make a difference.</li>\n<li>High motivation to learn and experiment with new technologies.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_07c95966-8e7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2589679","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Java","Kotlin","Spring Boot","Gradle","AWS","Kubernetes","ReactJS","EventBridge","SQS","ActiveMQ","PostgreSQL","S3","Valkey","ElasticSearch","GraphQL","OpenTelemetry","Grafana","Prometheus","ELK","APM","CloudWatch"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:06.987Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Kotlin, Spring Boot, Gradle, AWS, Kubernetes, ReactJS, EventBridge, SQS, ActiveMQ, PostgreSQL, S3, Valkey, ElasticSearch, GraphQL, OpenTelemetry, Grafana, Prometheus, ELK, APM, CloudWatch"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_60c33646-892"},"title":"Reference Data Operations Analyst","description":"<p>A Reference Data Operations Analyst is sought to join the Operations &amp; Middle Office department at Millennium. The successful candidate will partner with stakeholders across the firm to ensure accurate and timely setup of bank debt instruments in the Global Security Master. Key responsibilities include ensuring accurate creation and maintenance of all tradable securities, overseeing data quality control checks, creating and maintaining procedures and workflows, and liaising with internal and external clients.</p>\n<p>The ideal candidate will have a strong background in financial services, specifically in Bank Debt, Security Master and Reference Data. They will be a self-starter with close attention to detail, able to implement and maintain new systems, procedures, and controls, and possess strong communication and interpersonal skills.</p>\n<p>This is a full-time position with an estimated base salary range of $160,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_60c33646-892","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Middle Office","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954257548","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$160,000 to $250,000","x-skills-required":["Bloomberg","Refinitiv","Wall Street Office","ClearPar","Markit (S&P Global)","ICE","SQL queries","MS Excel"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:04.793Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"Bloomberg, Refinitiv, Wall Street Office, ClearPar, Markit (S&P Global), ICE, SQL queries, MS Excel","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_08e2a9b0-d54"},"title":"Software Engineer","description":"<p>As a Software Engineer at Equity IT, you will be part of the Latency-Critical Trading team, which is building a best-in-class systematic data platform to power the next generation of low-latency systematic strategies.</p>\n<p>The team includes low-latency Linux, network, datacenter, and C++ engineers focused on our end-to-end trading stack.</p>\n<p>Key responsibilities include monitoring and assessing the quality of live and historical market data, detecting, inventorying, and remedying data gaps, maintaining and documenting exchange session times, holiday schedules, timestamp rules, and protocol/microstructure changes, analyzing latency, data rates, bursts, and message flows to understand microstructure behaviour and system performance, cleaning, transforming, and managing an inventory of large-scale datasets, building and improving tools for market data capture, working with vendors and brokers to assess and provision datasets, building and improving tools for data analysis, visualization, and diagnostics on top of captured market and network data, enhancing and extending C++ analytics libraries and exposing them within a Python environment for systematic research and alpha development, and collaborating closely with portfolio managers, quantitative researchers, and engineers to translate trading use cases into robust data and tooling solutions.</p>\n<p>Qualifications include a Bachelor&#39;s or Master&#39;s in Computer Science, Mathematics, Statistics, Engineering, or another quantitative field, or equivalent experience, 3+ years of experience in financial markets, electronic trading, or high-frequency/systematic environments, strong programming skills in Python, C++, and SQL, solid understanding of modern statistical testing methods and comfort working with large, noisy, real-world datasets, experience with Linux, large-scale data processing, and preferably network data (PCAP, timestamping, PTP) and low-latency systems, and strong problem-solving skills, attention to detail, and effective communication with both technical and non-technical stakeholders.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_08e2a9b0-d54","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955295716","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","C++","SQL","Linux","large-scale data processing","network data (PCAP, timestamping, PTP)","low-latency systems"],"x-skills-preferred":["R","MATLAB","SciPy stack","PyTorch"],"datePosted":"2026-04-18T22:14:02.001Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, Karnataka, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, C++, SQL, Linux, large-scale data processing, network data (PCAP, timestamping, PTP), low-latency systems, R, MATLAB, SciPy stack, PyTorch"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d80e50b0-018"},"title":"Business Unit Control - Financing","description":"<p>We are seeking a highly skilled and motivated Business Unit Control - Financing professional to join our Operations &amp; Middle Office team in New York. This individual will play a critical role in the daily oversight, control, and reporting of financing P&amp;L across all asset classes, serving as a key liaison between Portfolio Managers, senior leadership, and cross-functional teams across the firm.</p>\n<p>This is a high-visibility role that sits at the intersection of finance, operations, and technology, offering the opportunity to work directly with some of the most sophisticated investment professionals in the industry.</p>\n<p>The ideal candidate is analytically rigorous, proactive, and thrives in a fast-paced, high-performance environment.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Daily P&amp;L Production &amp; Analysis , Own the end-to-end T+1 financing P&amp;L process, delivering timely and accurate P&amp;L explains, variance commentary, and attribution analysis across all asset classes.</li>\n<li>PM Interface , Serve as the primary point of contact for Portfolio Managers on all financing-related queries.</li>\n<li>Centralized Controls &amp; Validation , Act as a central control function by coordinating across Operations, Middle Office, Finance, and Technology teams to validate and reconcile all inputs used in the calculation of daily financing.</li>\n<li>Senior Management Reporting , Prepare and deliver high-quality daily and weekly reporting packages and analytics for the senior management team.</li>\n<li>Process Improvement &amp; Project Work , Drive the continuous enhancement of existing tools, workflows, and reporting infrastructure.</li>\n</ul>\n<p>Qualifications &amp; Skills Required:</p>\n<ul>\n<li>2+ years of relevant experience in Portfolio Finance, Product Control, or P&amp;L production within a financial services environment; hedge fund or investment bank experience strongly preferred.</li>\n<li>Proven track record of successfully interacting cross-functionally, specifically with investment professionals, trading desks, and senior management.</li>\n<li>Deep expertise across the following areas:</li>\n<li>Risk-Based P&amp;L Explained &amp; P&amp;L Attribution across multiple asset classes including equities, fixed income, derivatives, and FX.</li>\n<li>T+1 P&amp;L production , experience owning the full daily P&amp;L cycle from data collection through final delivery.</li>\n<li>Intraday and end-of-day P&amp;L support and reporting in a live trading environment.</li>\n<li>P&amp;L controls and reconciliations , ensuring accuracy, completeness, and consistency of all financial data across systems.</li>\n<li>General ledger and accounting data , ability to understand, reconcile, and articulate P&amp;L against the firm&#39;s books and records.</li>\n<li>Senior Management reporting , experience preparing executive-level deliverables that are accurate, insightful, and visually compelling.</li>\n<li>Automation and process improvement , demonstrated ability to identify operational inefficiencies and design and implement scalable solutions.</li>\n<li>Advanced Excel skills required , ability to build, maintain, and enhance complex financial models, pivot analyses, and reporting tools.</li>\n<li>Experience with Murex, Geneva, or Imagine is a strong plus.</li>\n<li>Proficiency in SQL for data extraction, validation, and analysis; Visio and PowerPoint skills preferred.</li>\n<li>Demonstrated initiative, intellectual curiosity, and a proactive problem-solving mindset , ability to identify issues before they escalate and drive solutions independently.</li>\n<li>Exceptional verbal and written communication skills, with the proven ability to present complex financial information clearly and confidently to individuals at all levels of the organization.</li>\n<li>Highly detail-oriented with strong ownership of work and an unwavering commitment to accuracy and quality under tight deadlines.</li>\n<li>Proven ability to prioritize and manage multiple tasks simultaneously in a fast-paced, deadline-driven environment while consistently delivering high-quality output.</li>\n<li>A collaborative team player with a strong desire to contribute to Millennium&#39;s continued reputation for operational excellence and investment performance.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d80e50b0-018","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Middle Office","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954991882","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$70,000 to $250,000","x-skills-required":["Risk-Based P&L Explained & P&L Attribution","T+1 P&L production","Intraday and end-of-day P&L support and reporting","P&L controls and reconciliations","General ledger and accounting data","Senior Management reporting","Automation and process improvement","Advanced Excel skills","Murex, Geneva, or Imagine","SQL","Visio and PowerPoint"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:59.352Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"Risk-Based P&L Explained & P&L Attribution, T+1 P&L production, Intraday and end-of-day P&L support and reporting, P&L controls and reconciliations, General ledger and accounting data, Senior Management reporting, Automation and process improvement, Advanced Excel skills, Murex, Geneva, or Imagine, SQL, Visio and PowerPoint","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_44a98c98-56e"},"title":"Business Unit Control - Loan Specialist","description":"<p>We are seeking a bright, motivated individual to join our Business Unit Control Group. As a Business Unit Control - Loan Specialist, you will be responsible for operational processing and full trade lifecycle management for financial reporting. You will review daily, monthly reconciliations independently and be able to escalate any issues.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Operational processing and full trade lifecycle management for financial reporting</li>\n<li>Review daily, monthly reconciliations independently, ability to be escalation point</li>\n<li>Experience with P&amp;L controls &amp; reconciliations; General ledger and accounting data</li>\n<li>Build and improve processes/workflows</li>\n<li>Partner with Finance for month-end close and audit</li>\n<li>Liaise across the organisation for information and problem solving</li>\n<li>Work with large data sets</li>\n<li>Ability to take trade and security data and transform into specs for technology to build from</li>\n</ul>\n<p>The ideal candidate will have 7+ years of relevant experience, a background in Accounting, Finance or Operations, and experience with a portfolio accounting systems, SS&amp;C Geneva.</p>\n<p>As a Business Unit Control - Loan Specialist, you will work closely with the Finance team to ensure accurate and timely financial reporting. You will also be responsible for building and improving processes and workflows to enhance efficiency and productivity.</p>\n<p>We offer a competitive total compensation package, including a base salary, discretionary performance bonus, and comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_44a98c98-56e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Middle Office","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954155762","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 to $250,000","x-skills-required":["P&L controls & reconciliations","General ledger and accounting data","Portfolio accounting systems","SS&C Geneva","Loan life cycle processes and events"],"x-skills-preferred":["Excel","SQL","Gresham reconciliation tool"],"datePosted":"2026-04-18T22:13:57.013Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"P&L controls & reconciliations, General ledger and accounting data, Portfolio accounting systems, SS&C Geneva, Loan life cycle processes and events, Excel, SQL, Gresham reconciliation tool","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b717f25a-5b3"},"title":"Data Operations Analyst - Systematic Data Platform","description":"<p>We are building a world-class systematic data platform that will power the next generation of our systematic portfolio engines.</p>\n<p>The systematic data group is looking for a Data Operations Analyst to join our growing team. The team consists of content specialists, data scientists, analysts, and engineers who are responsible for discovering, maintaining, and analysing sources of alpha for our portfolio managers.</p>\n<p>This is an opportunity for individuals who have a strong background in quantitative investing and are passionate about working with data.</p>\n<p>The role builds on individual&#39;s knowledge and skills in four key areas of quantitative investing: data, statistics, technology, and financial markets.</p>\n<p><strong>Principal Responsibilities</strong></p>\n<ul>\n<li>Efficiently monitor data flows across various systems, ensuring accuracy, completeness, and timeliness.</li>\n<li>Maintain and enhance the functionality and efficiency of our in-house data monitoring systems.</li>\n<li>Recommend and implement improvements to optimise data processing and quality.</li>\n<li>Design, build, and manage efficient and scalable data ingestion and ETL pipelines. Ensure smooth data flow from various sources into our core systems.</li>\n<li>Liaise with stakeholders across the organisation to understand their data requirements and support their initiatives.</li>\n<li>Actively engage with data issues in our production operation environment and aim to provide high-quality support on solving the issue both internally or with vendors.</li>\n</ul>\n<p><strong>Qualifications/Skills Required</strong></p>\n<ul>\n<li>Master&#39;s or Bachelor&#39;s in computer science, mathematics, statistics, or other field with good coding skills.</li>\n<li>2+ years of financial industry experience preferred.</li>\n<li>Programming expertise in Python, C++, Java, or C#.</li>\n<li>Programming skills in SQL, PL-SQL, or T-SQL.</li>\n<li>Strong problem-solving skills.</li>\n<li>Strong communication skills.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b717f25a-5b3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955700474","x-work-arrangement":null,"x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","C++","Java","SQL","PL-SQL","T-SQL"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:54.703Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hong Kong, Hong Kong"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Finance","skills":"Python, C++, Java, SQL, PL-SQL, T-SQL"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5c70414d-4e6"},"title":"Full‑Stack data engineer","description":"<p>We are seeking a highly self-sufficient, motivated engineer with strong full-stack data engineering skills to join our team. This is a remote/offshore role that requires autonomy, excellent communication, and the ability to deliver high-quality work with limited supervision while collaborating with a predominantly US-based team.</p>\n<p>You will build reliable, scalable data products and user experiences that power AI/ML modeling, agentic workflows, and reporting,working end-to-end from data ingestion and transformation through to UI. Our Python-based data platform is undergoing a major evolution toward a modern, cloud-native ELT architecture. We are standardizing on Snowflake as our central data platform and dbt as our core transformation framework, implementing scalable, maintainable ELT practices that simplify ingestion, modeling, and deployment.</p>\n<p>This role will be pivotal in independently designing and building robust data pipelines and semantic layers that directly power our AI and machine learning initiatives,delivering clean, reliable, and well-modeled data assets to our data science team for feature engineering, model training, and production inference. You will collaborate closely (primarily via remote channels) with data scientists and ML engineers to ensure our data ecosystem is optimized for experimentation speed, model performance, and seamless integration into downstream products and services.</p>\n<p>Key Responsibilities</p>\n<ul>\n<li>Remote collaboration &amp; communication: Operate effectively as an offshore member of a distributed team, proactively communicating status, risks, and blockers across time zones and coordinating overlap with US working hours as needed.</li>\n</ul>\n<ul>\n<li>Full-stack data engineering: Build across the entire stack, including data ingestion/acquisition and transformation, APIs, front-end components, and automated test suites, delivering production-grade solutions with minimal hand-holding.</li>\n</ul>\n<ul>\n<li>Autonomous delivery &amp; ownership: Take end-to-end ownership of features and projects,clarifying requirements, breaking work into milestones, estimating timelines, and delivering high-quality, well-documented solutions.</li>\n</ul>\n<ul>\n<li>Specification and design: Translate short- and long-term business requirements, architectural considerations, and competing timelines into clear, actionable technical specifications and design documents.</li>\n</ul>\n<ul>\n<li>Code quality: Write clean, maintainable, efficient code that adheres to evolving standards and quality processes, including unit tests and isolated integration tests in containerized environments.</li>\n</ul>\n<ul>\n<li>Continuous improvement: Contribute to agile practices and provide input on technical strategy, architectural decisions, and process improvements, continuously suggesting better tools, patterns, and automation.</li>\n</ul>\n<p>Required Skills &amp; Experience</p>\n<ul>\n<li>Professional experience: 5+ years in software engineering, with a full-stack background building complex, scalable data-engineering pipelines using data warehouse technology, SQL with dbt, Python, AWS with Terraform, and modern UI technologies.</li>\n</ul>\n<ul>\n<li>Modern data engineering: Strong experience with medallion data architecture patterns using data warehouse technologies (e.g., Snowflake), data transformation tooling (e.g., dbt), BI tooling, and NoSQL data marts (e.g., Elasticsearch/OpenSearch).</li>\n</ul>\n<ul>\n<li>Testing and QA: Solid understanding of unit testing, CI/CD automation, and quality assurance processes for both data pipeline testing and operational data quality tests.</li>\n</ul>\n<ul>\n<li>Remote work &amp; autonomy: Proven track record working in a remote or distributed environment, demonstrating self-motivation, reliable execution, and the ability to make sound technical decisions independently.</li>\n</ul>\n<ul>\n<li>Agile methodology: Working knowledge of Agile development practices and workflows (e.g., sprint planning, stand-ups, retrospectives) in a distributed team setting.</li>\n</ul>\n<ul>\n<li>Education: Bachelor’s or Master’s degree in Computer Science, Statistics, Informatics, Information Systems, or a related quantitative field.</li>\n</ul>\n<p>Preferred Skills &amp; Experience</p>\n<ul>\n<li>Machine learning and AI: Hands-on experience with large language models (LLMs) and agentic frameworks/workflows.</li>\n</ul>\n<ul>\n<li>Search and analytics: Familiarity with the ELK stack (Elasticsearch, Logstash, Kibana) for search and analytics solutions.</li>\n</ul>\n<ul>\n<li>Cloud expertise: Experience with AWS cloud services; familiarity with SageMaker; and CI/CD tooling such as GitHub Actions or Jenkins.</li>\n</ul>\n<ul>\n<li>Front-end expertise: Experience building user interfaces with Angular or a modern UI stack.</li>\n</ul>\n<ul>\n<li>Financial domain knowledge: Broad understanding of equities, fixed income, derivatives, futures, FX, and other financial instruments.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5c70414d-4e6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955321460","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Snowflake","dbt","AWS","Terraform","modern UI technologies","data warehouse technology","SQL","unit testing","CI/CD automation","quality assurance processes"],"x-skills-preferred":["machine learning","AI","large language models","agentic frameworks","ELK stack","search and analytics solutions","cloud expertise","AWS cloud services","SageMaker","CI/CD tooling","front-end expertise","Angular","financial domain knowledge"],"datePosted":"2026-04-18T22:13:54.584Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, Karnataka, India"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Snowflake, dbt, AWS, Terraform, modern UI technologies, data warehouse technology, SQL, unit testing, CI/CD automation, quality assurance processes, machine learning, AI, large language models, agentic frameworks, ELK stack, search and analytics solutions, cloud expertise, AWS cloud services, SageMaker, CI/CD tooling, front-end expertise, Angular, financial domain knowledge"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_87749959-700"},"title":"Intern Data Engineering (all genders)","description":"<p>Join our Data Engineering team inside the Business Intelligence department, where you&#39;ll work with experienced engineers to build the data foundation that powers Holidu&#39;s growth.</p>\n<p>As an intern, you&#39;ll get hands-on experience with real problems and have the opportunity to make a meaningful impact. You&#39;ll work on building and supporting data pipelines, digging into data quality, getting hands-on with cloud infrastructure, and exploring AI-assisted development.</p>\n<p>Our team uses a range of technologies, including Redshift, Athena, DuckDB, Terraform, Docker, Jenkins, ELK, Grafana, Looker, OpsGenie, Kafka, Airbyte, and Fivetran. You&#39;ll have the chance to learn from experienced engineers and contribute to the development of our data systems.</p>\n<p>In this role, you&#39;ll be part of a team that genuinely loves what they do and is passionate about building a better data foundation for Holidu. You&#39;ll have the opportunity to take responsibility from day one and develop through regular feedback.</p>\n<p>We offer a fair salary, the chance to make a difference for hundreds of thousands of monthly users, and the opportunity to grow and develop through regular feedback. You&#39;ll also have access to a range of benefits, including a hybrid work policy, the chance to work from other local offices, and a corporate subscription to Urban Sports Club or a premium gym membership at a discounted rate.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_87749959-700","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2557398","x-work-arrangement":"hybrid","x-experience-level":"intern","x-job-type":"Internship","x-salary-range":null,"x-skills-required":["Python","SQL","Git","Airflow","dbt","Docker","Cloud platform (AWS, GCP, etc.)"],"x-skills-preferred":["LLM tools","AI-assisted coding"],"datePosted":"2026-04-18T22:13:52.778Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Git, Airflow, dbt, Docker, Cloud platform (AWS, GCP, etc.), LLM tools, AI-assisted coding"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c3b63dd5-0f6"},"title":"Backend utvecklare","description":"<p>We are seeking an experienced backend developer to join our tech team. As a backend developer, you will be responsible for designing, developing, and maintaining the server-side of our applications and systems. You will work closely with our frontend developers, designers, and product owners to ensure a seamless integration between frontend and backend.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and develop scalable and efficient backend solutions for our digital platforms.</li>\n<li>Write clean, readable, and reusable code.</li>\n<li>Perform unit testing and debugging to ensure high quality and reliability.</li>\n<li>Participate in technical discussions and contribute ideas to improve the product&#39;s performance and functionality.</li>\n<li>Collaborate with frontend developers and other team members to ensure a smooth user experience.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Experience in backend development with a focus on web applications.</li>\n<li>Good knowledge of programming languages such as Python, Java, or similar.</li>\n<li>Experience working with frameworks such as Django, Flask, Spring, or similar.</li>\n<li>Familiarity with database management systems such as MySQL, PostgreSQL, or similar.</li>\n<li>Knowledge of API design and implementation.</li>\n<li>Strong problem-solving skills and ability to work independently as well as in a team.</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Attractive salary based on experience and competence.</li>\n<li>Opportunity to work with exciting projects and the latest technology.</li>\n<li>Flexible working hours and possibility of remote work.</li>\n<li>Continuous professional development and opportunities for career growth.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c3b63dd5-0f6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scandinavian Airlines","sameAs":"https://scandinavianairlines.teamtailor.com","logo":"https://logos.yubhub.co/scandinavianairlines.teamtailor.com.png"},"x-apply-url":"https://scandinavianairlines.teamtailor.com/jobs/4882026-backend-utvecklare","x-work-arrangement":"On-site","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["backend development","web applications","Python","Java","Django","Flask","Spring","MySQL","PostgreSQL","API design","problem-solving"],"x-skills-preferred":["cloud services","AWS","Google Cloud","Azure"],"datePosted":"2026-04-18T22:13:45.980Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Transportation","skills":"backend development, web applications, Python, Java, Django, Flask, Spring, MySQL, PostgreSQL, API design, problem-solving, cloud services, AWS, Google Cloud, Azure"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ad717304-da7"},"title":"Intern Data Analytics (all genders)","description":"<p>You will be part of the Business Intelligence department, which consists of the Data Science, Data Analytics, and Data Engineering teams.</p>\n<p>This internship provides a great opportunity to gain hands-on experience into Data Analytics. You will work alongside a team of highly skilled and dedicated professionals who are committed to offering strong mentorship and guidance to help you start your career in the field of data.</p>\n<p>Duration: 6 months. Location: Munich, 2-3 office days per week.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git, Atlassian.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>As a Data Analytics Intern at Holidu, you’ll help our company make smarter, data-driven decisions, while being supported by a Senior Analyst.</p>\n<p>This role goes beyond building dashboards. We want curious, proactive people who want to become data advisors - not only delivering reports, but understanding the business context, which questions they answer and why they matter.</p>\n<ul>\n<li>Collect, analyse, and interpret large datasets to help solve real business challenges.</li>\n<li>Build dashboards and reports using tools like SQL, Python, and Looker.</li>\n<li>Collaborate closely with teams such as Product, Marketing, or Finance to help them extract actionable insights from data.</li>\n<li>Build and improve data pipelines using cutting-edge technologies.</li>\n<li>We are an AI-first team. Rather than manually executing repetitive tasks, you will use AI to work smarter and automate workflows.</li>\n<li>You’ll collaborate with our Data Scientists and get exposure to:</li>\n<li>Data preparation and exploratory data analysis.</li>\n<li>How ML-models are built, evaluated, and deployed in real-life.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>Currently enrolled in or recently completed a Bachelor’s or Master’s degree in a quantitative field (e.g., Business Analytics, Data Science, Economics, Statistics, Mathematics, Engineering or similar).</li>\n<li>Understanding of SQL and Python, proficiency in Excel/Google Sheets and a desire to learn visualization tools like Looker.</li>\n<li>Knowledge of Machine Learning and Statistical models is a plus.</li>\n<li>Strong analytical and problem-solving skills, and attention to detail.</li>\n<li>Curiosity to learn and a passion for solving data problems.</li>\n<li>Good communication and presentation skills.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Compensation: Get a fair salary.</li>\n<li>Impact: Make a difference for hundreds of thousands of monthly users.</li>\n<li>Growth: Take responsibility from day one and develop through regular feedback.</li>\n<li>Community: Engage with international, diverse, yet like-minded colleagues through regular events and 2 office days per week with your team.</li>\n<li>Flexibility: Benefit from our hybrid work policy and the chance to work from other local offices for up to 8 weeks a year.</li>\n<li>Fitness: Get a Urban Sports Club corporate subscription or a premium gym membership at a discounted rate.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ad717304-da7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2556233","x-work-arrangement":"hybrid","x-experience-level":"intern","x-job-type":"Internship","x-salary-range":null,"x-skills-required":["SQL","Python","Looker","Git","Atlassian","Airflow","DBT","AWS Stack","Redshift","Athena","Glue","S3"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:45.423Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Looker, Git, Atlassian, Airflow, DBT, AWS Stack, Redshift, Athena, Glue, S3"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_cc9213ff-135"},"title":"(Senior) Team Lead Marketing Analytics (all genders)","description":"<p>Within the Marketing Technology department, we are building a new Marketing Analytics team and are looking for a Team Lead to shape it from the ground up.</p>\n<p>You&#39;ll work closely with a wide range of Marketing stakeholders, ensuring they have the data, tools, and insights they need to drive sustainable growth. Moreover, you will also collaborate with data scientists and data engineers within the department to build best-in-class analytical solutions.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git, Jira, Confluence, Slack.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<ul>\n<li>You&#39;ll be leading data analysts and collaborating cross-functionally with data engineers and data scientists - fostering collaboration, learning, and analytical excellence.</li>\n<li>Engage with senior marketing leadership on strategic projects, providing insights that influence channel strategy and budget decisions, and ultimately our revenue growth.</li>\n<li>Translate marketing logic, for a diverse range of channels (e.g. Performance Marketing, SEO, CRM, affiliate) and use cases into analytical requirements and communicate complex findings clearly to both technical and commercial teams.</li>\n<li>Support and partner with Marketing Technology on tracking, event design, and data flows to ensure data quality and reliable reporting frameworks.</li>\n<li>Not shying away from hands-on work as an individual contributor (50% to start), while leading the team, diving deep into the details when needed.</li>\n<li>Shape the future of marketing analytics at Holidu by recruiting top talent, setting clear goals, and developing your team personally and professionally.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>5+ years multi-channel marketing analytics experience in a B2B or B2C organisation where marketing is a core performance driver, with extensive hands-on expertise in at least one of the following: attribution, cost and revenue allocation, or bidding.</li>\n<li>People management experience - this should not be your first leadership role.</li>\n<li>A collaborative mindset with clear experience communicating with executive stakeholders and senior decision makers.</li>\n<li>You are mission-driven, with a working backwards mentality (i.e. starting with customer needs) and clear experience managing and delivering complex projects with multiple stakeholders. Ability to translate business goals into analytical solutions and break down complex topics into actionable insights.</li>\n<li>Excellent analytical and technical skills. Concretely: strong in SQL, Python (or similar), data visualisation skills as well as developing technical frameworks to serve a clear business need.</li>\n<li>A strong personal or team focus on AI enablement: you actively use AI tools to enhance your coding, planning, and workflows, and can enable your team to do the same.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_cc9213ff-135","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2458940","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["AWS Stack","Airflow","DBT","Looker","SQL","Python","Git","Jira","Confluence","Slack"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:45.213Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Marketing","industry":"Technology","skills":"AWS Stack, Airflow, DBT, Looker, SQL, Python, Git, Jira, Confluence, Slack"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_32932504-2b5"},"title":"Systematic Production Support Engineer","description":"<p>We are looking for an experienced professional to help us scale our systematic operations and support engineering capabilities.</p>\n<p>This role directly supports portfolio management teams across Millennium, with operational excellence at the core. Our efforts are focused on delivering the highest quality returns to our investors – providing a world-class and reliable trading and technology platform is essential to this mission.</p>\n<p>This is a unique opportunity to drive significant value creation for one of the world&#39;s leading investment managers.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Build, develop and maintain a reliable, scalable, and integrated platform for trading strategy monitoring, reporting, and operations.</li>\n<li>Work with portfolio managers and other internal customers to reduce operational risk through:</li>\n<li>Implementation of monitoring, reporting, and trade workflow solutions.</li>\n<li>Implementation of automated systems and processes focused on trading and operations.</li>\n<li>Streamlining development and deployment processes.</li>\n<li>Implementation of MCP servers focused on assisting rest of the Support Engineering team as well as proactively monitoring production environment.</li>\n</ul>\n<p>Technical Qualification:</p>\n<ul>\n<li>5+ years of development experience in Python.</li>\n<li>Experience working in a Linux / Unix environment.</li>\n<li>Experience working with PostgreSQL or other relational databases.</li>\n<li>Ability to understand and discuss requirements from portfolio managers.</li>\n</ul>\n<p>Preferred Skills and Experience:</p>\n<ul>\n<li>Understanding of NLP, supervised/non-supervised learning and Generative AI models.</li>\n<li>Experience operating and monitoring low-latency trading environments.</li>\n<li>Familiarity with quantitative finance and electronic trading concepts.</li>\n<li>Familiarity with financial data.</li>\n<li>Broad understanding of equities, futures, FX, or other financial instruments.</li>\n<li>Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#.</li>\n<li>Experience with Apache / Confluent Kafka.</li>\n<li>Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline).</li>\n<li>Experience with containerization and orchestration technologies.</li>\n<li>Experience building and deploying systems that utilize services provided by AWS, GCP or Azure.</li>\n<li>Contributions to open-source projects.</li>\n</ul>\n<p>The estimated base salary range for this position is $100,000 to $175,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package. When finalizing an offer, we take into consideration an individual&#39;s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_32932504-2b5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954627501","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$100,000 to $175,000","x-skills-required":["Python","Linux / Unix","PostgreSQL","NLP","supervised/non-supervised learning","Generative AI models"],"x-skills-preferred":["Apache / Confluent Kafka","C/C++","Java","Scala","Go","C#","containerization","orchestration technologies","AWS","GCP","Azure"],"datePosted":"2026-04-18T22:13:42.254Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America · Old Greenwich, Connecticut, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Linux / Unix, PostgreSQL, NLP, supervised/non-supervised learning, Generative AI models, Apache / Confluent Kafka, C/C++, Java, Scala, Go, C#, containerization, orchestration technologies, AWS, GCP, Azure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":100000,"maxValue":175000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_af78786b-a0a"},"title":"Software Engineer - Compliance / Regulatory Reporting","description":"<p>The Compliance/Regulatory Reporting technology team at Millennium builds solutions to meet the firm&#39;s global regulatory and reporting obligations.</p>\n<p>We use AI-assisted development tools (e.g., Claude Code), cloud-native/serverless architectures on AWS, and modern full-stack technologies (C#, Angular, SQL), with a strong focus on Domain-Driven Design (DDD) and automated testing.</p>\n<p>The role is suited to engineers who have delivered real-time, mission-critical systems in high trade volume, distributed and fault-tolerant environments.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and build scalable, real-time Regulatory/Compliance applications using C#/.NET, Angular, and SQL, leveraging AI-assisted tools to accelerate development and improve quality.</li>\n<li>Model business domains using DDD (bounded contexts, aggregates, entities, value objects, domain services, domain events) with a strong focus on business correctness and ubiquitous language.</li>\n<li>Architect and implement cloud-native/serverless solutions on AWS, including:</li>\n<li>Event-driven services using AWS Lambda and messaging/streaming (Kafka, SQS, SNS).</li>\n<li>Containerized microservices using Docker and Kubernetes (e.g., Amazon EKS).</li>\n<li>Build and maintain Angular front-ends that integrate securely and efficiently with backend APIs and domain services.</li>\n<li>Design and optimize relational data models and SQL queries (SQL Server, Snowflake) for high-volume, low-latency workloads.</li>\n<li>Drive a test-first mindset with strong automated test coverage (unit, integration, contract, and end-to-end) for critical domain workflows and controls.</li>\n<li>Collaborate with global business and Compliance stakeholders to understand requirements, shape domain models, and deliver auditable, production-ready solutions.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Core Engineering &amp; Full-Stack Skills</li>\n</ul>\n<p>+ Practical experience with AI-assisted tools (e.g., Claude Code, GitHub Copilot) for code generation/refactoring, test creation, debugging, and documentation   + Expert-level C#/.NET and strong object-oriented design skills   + Solid experience building Angular applications (components, state, routing, API integration)   + Advanced SQL skills for schema design and complex queries (SQL Server, Snowflake)   + Experience with high-throughput, concurrent/multithreaded systems   + Kafka or similar messaging experience, including using JSON and Avro for data contracts in streaming and messaging   + Strong understanding of unit testing, Dependency Injection, design patterns, concurrency, and SOLID principles   + Experience with Git and GitHub in a collaborative, code-review-driven workflow</p>\n<ul>\n<li>Soft Skills &amp; Domain Knowledge</li>\n</ul>\n<p>+ Excellent analytical and problem-solving abilities.   + Self-starter who thrives in a fast-paced, globally distributed environment.   + Strong written and verbal communication skills with the ability to explain domain models, testing strategies, and architectural decisions to varied audiences.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_af78786b-a0a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955321458","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["AI-assisted tools","C#/.NET","Angular","SQL","Domain-Driven Design","AWS","Kafka","Git","GitHub"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:41.349Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore, Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"AI-assisted tools, C#/.NET, Angular, SQL, Domain-Driven Design, AWS, Kafka, Git, GitHub"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6690b2fa-cab"},"title":"(Senior) Team Lead Data Analytics (all genders)","description":"<p>At Holidu, data isn&#39;t just a support function, it&#39;s how we make decisions. The Analytics team builds the products and foundations that keep the whole organisation sharp, from day-to-day operations to long-term strategy.</p>\n<p>This role is on-site in Munich, with two office days per week.</p>\n<p>As a Senior Team Lead Data Analytics, you will lead one of Holidu&#39;s core analytics teams, a function at the intersection of data, strategy, and real business impact. The team has four direct reports and entails collaborating cross-functionally with data engineers and data scientists.</p>\n<p>Engage with senior leadership on strategic projects, providing insights that influence product strategy, internal operations, and revenue growth.</p>\n<p>You and your team will support a range of stakeholders across the company (e.g. Customer Support, Host Experience, Sales and Account Management).</p>\n<p>As a member of the BI leadership team, you will help shape the department strategy and the future of AI-powered data products.</p>\n<p>Understand problems and identify opportunities across a diverse range of stakeholder use cases, translating them into analytical requirements and communicating complex findings clearly to both technical and commercial audiences.</p>\n<p>Lead from the front: this role carries meaningful individual contributor responsibility. You&#39;ll be expected to do real analytical work, diving deep into the data, building solutions, and setting the bar for quality in your team.</p>\n<p>Shape the future of analytics at Holidu by recruiting top talent, setting clear goals, and developing your team personally and professionally.</p>\n<p>The ideal candidate will have 5+ years of data analytics experience, people management experience, a collaborative mindset, a mission-driven mentality, excellent analytical and technical skills, and a genuine commitment to AI enablement.</p>\n<p>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</p>\n<p>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</p>\n<p>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</p>\n<p>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</p>\n<p>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</p>\n<p>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6690b2fa-cab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2598226","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Database: AWS Stack (Redshift, Athena, Glue, S3)","Data Pipelines: Airflow, dbt","Data Visualisation: Looker","Data Analytics: SQL, Python","Collaboration: Git, Jira, Confluence, Slack"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:28.264Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Technology","industry":"Travel Technology","skills":"Database: AWS Stack (Redshift, Athena, Glue, S3), Data Pipelines: Airflow, dbt, Data Visualisation: Looker, Data Analytics: SQL, Python, Collaboration: Git, Jira, Confluence, Slack"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f992fe0e-8ea"},"title":"C++ Software Engineer - Trade Capture","description":"<p>We are seeking a skilled C++ software engineer to join our global team in developing a real-time trade capture system. The successful candidate will have strong experience in C++ server-side applications, Linux, and real-time/event-driven systems. They will work closely with our team to design, develop, and implement features that meet the evolving needs of our clients.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Develop and maintain high-quality C++ code for our trade capture system</li>\n<li>Collaborate with cross-functional teams to identify and prioritize feature requirements</li>\n<li>Design and implement efficient algorithms and data structures to meet system performance goals</li>\n<li>Troubleshoot and resolve complex technical issues in a timely manner</li>\n<li>Participate in code reviews and contribute to the improvement of our codebase</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science or a related field</li>\n<li>7+ years of experience in developing C++ server-side applications in Linux</li>\n<li>Strong understanding of network programming and Linux O/S internals</li>\n<li>Strong understanding of object-oriented programming and data structures</li>\n<li>Understanding of SQL and relational databases</li>\n<li>Knowledge in Equities, Futures, FX, Options, and various Fixed Income products is strongly preferred</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Competitive total compensation package including base salary, discretionary performance bonus, and comprehensive benefits</li>\n<li>Opportunity to work with a talented team of professionals in a dynamic and fast-paced environment</li>\n<li>Flexible working hours and remote work options</li>\n</ul>\n<p>When finalizing an offer, we take into consideration an individual&#39;s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f992fe0e-8ea","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755943471581","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 to $250,000","x-skills-required":["C++","Linux","real-time/event-driven systems","network programming","object-oriented programming","data structures","SQL","relational databases"],"x-skills-preferred":["Equities","Futures","FX","Options","Fixed Income products"],"datePosted":"2026-04-18T22:13:22.371Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"C++, Linux, real-time/event-driven systems, network programming, object-oriented programming, data structures, SQL, relational databases, Equities, Futures, FX, Options, Fixed Income products","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_107cbb3f-b6c"},"title":"Production Support Engineer","description":"<p>The Production Support Engineer role is a hands-on, business-facing position that requires understanding how applications support the business, investigating functional and data-related issues, and communicating clearly with users under pressure.</p>\n<p>The Core Technology Production Support team supports a suite of business-critical financial applications used by Middle Office, Operations, Treasury, and Trading. These platforms are central to the firm&#39;s PnL, risk, cash, trade processing, and regulatory reporting workflows.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>End to end ownership of the production environment</li>\n<li>Infrastructure management</li>\n<li>Release planning and deployment</li>\n<li>Incident and problem management, including root cause analysis</li>\n<li>Capacity Planning / BCP Testing</li>\n<li>Build strong relationships with development and end-users/clients</li>\n<li>Foster the DevOps culture</li>\n<li>Focus on client service and delivery</li>\n<li>Become the go-to person for your area of responsibility</li>\n<li>Build subject matter expertise</li>\n<li>Create and maintain high quality documentation and runbooks</li>\n<li>Cross train other Support team members</li>\n</ul>\n<p>Qualifications/Skills Required:</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Electrical Engineering, or a related field.</li>\n<li>Minimum 2+ years’ experience supporting an enterprise environment</li>\n<li>Must have previous experience supporting business facing applications</li>\n<li>Strong scripting skills in one of the following: Python (preferred), PowerShell, Perl, etc.</li>\n<li>Excellent SQL skills and knowledge of various database systems</li>\n<li>Must be able to run and understand complex queries</li>\n<li>Ability to support both Windows and Unix/Linux environments</li>\n</ul>\n<p>Preferred Skills:</p>\n<ul>\n<li>Experience working in a trading environment</li>\n<li>Exposure to the following:</li>\n</ul>\n<ul>\n<li>CI/CD (Jenkins/Octopus/Artifactory)</li>\n<li>Metrics/KPIs (Datadog/Influx/Tableau)</li>\n<li>Kafka</li>\n<li>Kubernetes</li>\n<li>AI (MCP/Agents)</li>\n</ul>\n<p>The estimated base salary range for this position is $100,000 to $175,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_107cbb3f-b6c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755943534669","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$100,000 to $175,000","x-skills-required":["Python","PowerShell","Perl","SQL","Windows","Unix/Linux"],"x-skills-preferred":["CI/CD","Metrics/KPIs","Kafka","Kubernetes","AI"],"datePosted":"2026-04-18T22:13:14.556Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Finance","skills":"Python, PowerShell, Perl, SQL, Windows, Unix/Linux, CI/CD, Metrics/KPIs, Kafka, Kubernetes, AI","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":100000,"maxValue":175000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_25fd58ed-3c0"},"title":"(Senior) Data Scientist (all genders)","description":"<p>You will be part of the Business Intelligence department, which consists of the Data Science, Data Analytics, and Data Engineering teams.</p>\n<p>As a Senior Data Scientist, you will work on various topics such as rankings, recommendations, user segmentation, user lifetime value, business forecasts, etc. You will have access to our huge dataset and work in collaboration with stakeholders from various departments.</p>\n<p>Your objective is to build the best internal and external products for our customers. Holidu highly values a diverse and open environment with people from all over the world.</p>\n<p>This role is based in Munich with a hybrid setup.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Flexible data science environment (Python, Sagemaker)</li>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>You will play a pivotal role in the Business Intelligence team alongside data scientists, analysts, and engineers. Together, you will lead the development and enhancement of our company-wide machine learning strategy.</p>\n<ul>\n<li>Collaborate across various business departments to identify opportunities and solve critical business challenges using data science solutions.</li>\n<li>Build and optimize predictive models such as booking cancellation forecasts, churn predictions, pricing optimization, revenue forecasting and marketing channel allocation.</li>\n<li>Take models from conception to production, continuously monitor their performance, and iterate to enhance accuracy and efficiency.</li>\n<li>Interface with diverse business stakeholders, ensuring alignment between data science initiatives and company goals.</li>\n<li>Demonstrate leadership in data science projects, leveraging your expertise to drive measurable business impact.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>3+ years of experience as a Data Scientist, with a proven track record of applying data science methodologies to solve complex business problems.</li>\n<li>A degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field.</li>\n<li>Expertise in statistics, predictive analytics, machine learning techniques, and proficiency in tools like Python and SQL.</li>\n<li>Experience with Airflow and dbt is a plus.</li>\n<li>Strong understanding of business operations and experience collaborating with diverse stakeholders.</li>\n<li>Enthusiasm for data science and a drive to deliver world-class products that make a difference.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other.</li>\n<li>Technology: Work in a modern tech environment.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_25fd58ed-3c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2555141","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","Sagemaker","AWS Stack","Airflow","DBT","Looker","SQL","Git"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:07.588Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Sagemaker, AWS Stack, Airflow, DBT, Looker, SQL, Git"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5dfa9c86-5c0"},"title":"Director, US Forecasting & Analytics– Vaccines & Immune Therapies","description":"<p>Director, US Forecasting &amp; Analytics – Vaccines &amp; Immune Therapies Global Insights, Analytics &amp; Forecasting, BBU Hybrid Work- on average 3 days a week from office</p>\n<p>The Director, US Forecasting &amp; Analytics – Vaccines &amp; Immune Therapies is a senior commercial insights leader responsible for US demand forecasting and analytics across the V&amp;I portfolio. The role is predominantly forecast-focused, serving as the US forecasting lead and strategic thought partner to Marketing, Finance, Market Access, and US teams.</p>\n<p>Responsibilities:</p>\n<p>US Forecasting Leadership (Core Accountability)</p>\n<ul>\n<li>Lead US short-term and long-term demand forecasts (TRx, NBRx, volume, patients, revenue) for V&amp;I assets using robust, patient-based and market-based models</li>\n<li>Own forecast methodology, assumptions, and governance, ensuring objectivity, transparency, and consistency with enterprise standards</li>\n<li>Integrate primary market research, epidemiology, competitive intelligence, access dynamics, and real-world data into forecast models</li>\n<li>Proactively identify and quantify key risks and opportunities through scenario and sensitivity analyses</li>\n<li>Partner closely with Finance, Market Access &amp; Pricing, Marketing, Sales, Medical, and Global Forecasting to ensure alignment on assumptions and implications</li>\n<li>Support business planning, governance reviews, and opportunity assessments with clear, executive-ready narratives</li>\n<li>Serve as a trusted advisor to senior marketing and finance leadership, clearly articulating forecast drivers and changes</li>\n</ul>\n<p>Analytics &amp; Resource Leadership (Enablement)</p>\n<ul>\n<li>Provide leadership over forecasting-adjacent analytics, ensuring advanced analytics and insights are embedded into forecasting and business planning</li>\n<li>Manage and prioritize internal analysts, contractors, and external vendors supporting forecasting and analytics deliverables</li>\n<li>Partner with data analytics resources, Global IA&amp;F, and GIBEX capability teams to deploy new tools, data sources, and modeling approaches</li>\n<li>Champion and identify new ways to embed AI and advanced automation into the practice of data analytics and forecasting to drive efficiency, scalability, and decision quality</li>\n<li>Champion continuous improvement in forecasting processes, AI-enabled modeling, and automation</li>\n<li>Contribute to the development and sharing of best practices across the V&amp;I forecasting community</li>\n</ul>\n<p>Essential for the role</p>\n<ul>\n<li>Bachelor’s degree in a quantitative, scientific, or business-related field required (e.g., Statistics, Economics, Mathematics, Engineering, Computer/Data Science).</li>\n<li>8+ years’ experience in US pharmaceutical commercial forecasting, including in-market and late-stage pipeline assets</li>\n<li>Hands-on model ownership experience (build, refresh, and performance tracking) across short- and long-term horizons</li>\n<li>Expertise in scenario-based forecasting, sensitivity analysis, and driver-based narratives to support senior decision-making</li>\n<li>Strong capability integrating multiple data types (e.g., IQVIA, claims, epidemiology, RWD/RWE, primary research) into coherent, decision-grade forecasts</li>\n<li>Working knowledge of advanced analytics/ML approaches (e.g., time series, causal inference, ensembles) and where they add value vs. traditional methods</li>\n<li>Fluency in modern analytics tooling and automation (e.g., Python/R/SQL, BI/visualization), with ability to partner effectively with data engineering and analytics teams</li>\n<li>Demonstrated forecast governance and model risk discipline (traceable assumptions, documentation, and clear explanations)</li>\n<li>Strong understanding of US market access and payer dynamics and how they impact demand (coverage, contracting, channel, policy)</li>\n<li>Exceptional communication: translates complex analysis into clear, executive-ready insights, options, and recommendations</li>\n<li>Strong commercial competence across key demand levers (positioning, adoption, competitive dynamics, lifecycle events)</li>\n</ul>\n<p>Desirable for the role</p>\n<ul>\n<li>Advanced degree preferred (e.g., MBA, MS, PhD in Statistics, Economics, Decision Sciences, Data Science, or related discipline).</li>\n<li>Vaccines and/or Rare Disease experience, including familiarity with immunization dynamics, patient-based forecasting, and lifecycle management in preventive or immune-mediated therapies</li>\n<li>Change leadership: builds adoption for new tools, processes, and ways of working across cross-functional stakeholders</li>\n<li>Product mindset for forecasting: defines user needs, success metrics, and a roadmap for portfolio forecasting capabilities</li>\n<li>Model lifecycle practices (e.g., reproducibility, versioning, monitoring/drift awareness); familiarity with MLOps concepts</li>\n</ul>\n<p>Office Working Requirements</p>\n<p>When we put unexpected teams in the same room, we unleash bold thinking with the power to inspire life-changing medicines. In-person working gives us the platform we need to connect, work at pace and challenge perceptions. That’s why we work, on average, a minimum of three days per week from the office. But that doesn’t mean we’re not flexible. We balance the expectation of being in the office while respecting individual flexibility. Join us in our unique and ambitious world.</p>\n<p>#LI-Hybrid</p>\n<p>Date Posted 10-Apr-2026 Closing Date 23-Apr-2026 Our mission is to build an inclusive environment where equal employment opportunities are available to all applicants and employees. In furtherance of that mission, we welcome and consider applications from all qualified candidates, regardless of their protected characteristics. If you have a disability or special need that requires accommodation, please complete the corresponding section in the application form.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5dfa9c86-5c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Global Insights, Analytics & Forecasting - V&I","sameAs":"https://astrazeneca.eightfold.ai","logo":"https://logos.yubhub.co/astrazeneca.eightfold.ai.png"},"x-apply-url":"https://astrazeneca.eightfold.ai/careers/job/563877689756206","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["forecasting","analytics","model ownership","scenario-based forecasting","sensitivity analysis","driver-based narratives","advanced analytics","machine learning","Python","R","SQL","BI/visualization","data engineering","forecast governance","model risk discipline","US market access","payer dynamics","exceptional communication","commercial competence"],"x-skills-preferred":["vaccines","rare disease","change leadership","product mindset","model lifecycle practices","MLOps"],"datePosted":"2026-04-18T22:13:06.502Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Wilmington, Delaware, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Healthcare","skills":"forecasting, analytics, model ownership, scenario-based forecasting, sensitivity analysis, driver-based narratives, advanced analytics, machine learning, Python, R, SQL, BI/visualization, data engineering, forecast governance, model risk discipline, US market access, payer dynamics, exceptional communication, commercial competence, vaccines, rare disease, change leadership, product mindset, model lifecycle practices, MLOps"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_867e3558-9a7"},"title":"Team Lead, Java Engineer - Equities Trading Technologies","description":"<p>We are seeking a Team Lead to maintain and enhance our mission-critical, multi-asset trading platform that is used firm-wide daily. This individual will own the existing Java Swing code base, while also playing a pivotal role in designing the next-generation HTML5 trading UI.</p>\n<p>The ideal candidate should have a proven track record in developing and maintaining Java-based front-end applications in the finance sector. Exceptional team collaboration skills and the ability to work effectively with colleagues across global time zones are crucial.</p>\n<p>Millennium strongly prioritizes our synergistic culture, which revolves around teamwork and low egos. You should possess the ability to work in a fast-paced environment both collaboratively and individually while managing multiple projects simultaneously.</p>\n<p>The successful individual will have a strong sense of urgency, emotional intelligence, and prioritize a high-caliber end-user experience.</p>\n<p>Qualifications:</p>\n<ul>\n<li>Bachelor’s degree in computer science or comparable</li>\n<li>7+ years of professional experience with Core Java and Java Swing, electronic trading systems and/or trader workstations environment strongly preferred.</li>\n<li>5+ years of experience working with HTML, JavaScript, CSS, and JQuery</li>\n<li>Deep understanding of multithreading and distributed systems within a high performance, latency-sensitive environment</li>\n<li>Strong knowledge of unit testing frameworks and continuous test-driven development practices</li>\n<li>Enterprise level experience with design patterns such as MVC, MV, MVP</li>\n<li>Enterprise level experience with RESTful web services</li>\n<li>Previous experience liaising with non-technology stakeholders, polished and proactive communication skills</li>\n</ul>\n<p>Beneficial/Ideal Technology Experience:</p>\n<ul>\n<li>EXT-JS, AngularJS, AJAX, JSON experience is very beneficial</li>\n<li>Knowledge of equities, futures, options and other asset classes is preferred</li>\n<li>Enterprise level experience with OMS architecture and design is preferred</li>\n<li>Experience with messaging middleware, Solace preferred</li>\n<li>Experience with relational and NoSQL databases. MongoDB preferred</li>\n<li>Experience working with financial data, including reference data, market data, order/execution and positions data.</li>\n<li>Experience working with Cloud: AWS (preferred), GCP or Azure</li>\n</ul>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package. The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_867e3558-9a7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955412056","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Core Java","Java Swing","HTML","JavaScript","CSS","JQuery","Multithreading","Distributed systems","Unit testing frameworks","Continuous test-driven development practices","MVC","MV","MVP","RESTful web services"],"x-skills-preferred":["EXT-JS","AngularJS","AJAX","JSON","Equities","Futures","Options","OMS architecture and design","Messaging middleware","Solace","Relational databases","NoSQL databases","MongoDB","Financial data","Cloud","AWS","GCP","Azure"],"datePosted":"2026-04-18T22:13:00.318Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America · New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Core Java, Java Swing, HTML, JavaScript, CSS, JQuery, Multithreading, Distributed systems, Unit testing frameworks, Continuous test-driven development practices, MVC, MV, MVP, RESTful web services, EXT-JS, AngularJS, AJAX, JSON, Equities, Futures, Options, OMS architecture and design, Messaging middleware, Solace, Relational databases, NoSQL databases, MongoDB, Financial data, Cloud, AWS, GCP, Azure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7275ef33-009"},"title":"Staff Data Engineer","description":"<p>At Bayer, we&#39;re seeking a Staff Data Engineer to join our team. As a Staff Data Engineer, you will design and lead the implementation of data flows to connect operational systems, data for analytics and business intelligence (BI) systems. You will recognize opportunities to reuse existing data flows, lead the build of data streaming systems, optimize the code to ensure processes perform optimally, and lead work on database management.</p>\n<p>Communicating Between Technical and Non-Technical Colleagues</p>\n<p>As a Staff Data Engineer, you will communicate effectively with technical and non-technical stakeholders, support and host discussions within a multidisciplinary team, and be an advocate for the team externally.</p>\n<p>Data Analysis and Synthesis</p>\n<p>You will undertake data profiling and source system analysis, present clear insights to colleagues to support the end use of the data.</p>\n<p>Data Development Process</p>\n<p>You will design, build and test data products that are complex or large scale, build teams to complete data integration services.</p>\n<p>Data Innovation</p>\n<p>You will understand the impact on the organization of emerging trends in data tools, analysis techniques and data usage.</p>\n<p>Data Integration Design</p>\n<p>You will select and implement the appropriate technologies to deliver resilient, scalable and future-proofed data solutions and integration pipelines.</p>\n<p>Data Modeling</p>\n<p>You will produce relevant data models across multiple subject areas, explain which models to use for which purpose, understand industry-recognised data modelling patterns and standards, and when to apply them, compare and align different data models.</p>\n<p>Metadata Management</p>\n<p>You will design an appropriate metadata repository and present changes to existing metadata repositories, understand a range of tools for storing and working with metadata, provide oversight and advice to more inexperienced members of the team.</p>\n<p>Problem Resolution</p>\n<p>You will respond to problems in databases, data processes, data products and services as they occur, initiate actions, monitor services and identify trends to resolve problems, determine the appropriate remedy and assist with its implementation, and with preventative measures.</p>\n<p>Programming and Build</p>\n<p>You will use agreed standards and tools to design, code, test, correct and document moderate-to-complex programs and scripts from agreed specifications and subsequent iterations, collaborate with others to review specifications where appropriate.</p>\n<p>Technical Understanding</p>\n<p>You will understand the core technical concepts related to the role, and apply them with guidance.</p>\n<p>Testing</p>\n<p>You will review requirements and specifications, and define test conditions, identify issues and risks associated with work, analyse and report test activities and results.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7275ef33-009","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976928777","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$114,400 to $171,600","x-skills-required":["Proficiency in programming language such as Python or Java","Experience with Big Data technologies such as Hadoop, Spark, and Kafka","Familiarity with ETL processes and tools","Knowledge of SQL and NoSQL databases","Strong understanding of relational databases","Experience with data warehousing solutions","Proficiency with cloud platforms","Expertise in data modeling and design","Experience in designing and building scalable data pipelines","Experience with RESTful APIs and data integration"],"x-skills-preferred":["Relevant certifications (e.g., GCP Certified, AWS Certified, Azure Certified)","Bachelor's degree in Computer Science, Data Engineering, Information Technology, or a related field","Strong analytical and communication skills","Ability to work collaboratively in a team environment","High level of accuracy and attention to detail"],"datePosted":"2026-04-18T22:12:56.654Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"Proficiency in programming language such as Python or Java, Experience with Big Data technologies such as Hadoop, Spark, and Kafka, Familiarity with ETL processes and tools, Knowledge of SQL and NoSQL databases, Strong understanding of relational databases, Experience with data warehousing solutions, Proficiency with cloud platforms, Expertise in data modeling and design, Experience in designing and building scalable data pipelines, Experience with RESTful APIs and data integration, Relevant certifications (e.g., GCP Certified, AWS Certified, Azure Certified), Bachelor's degree in Computer Science, Data Engineering, Information Technology, or a related field, Strong analytical and communication skills, Ability to work collaboratively in a team environment, High level of accuracy and attention to detail","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":114400,"maxValue":171600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_37654a6b-cb7"},"title":"Senior Clinical Data Manager II","description":"<p>At AstraZeneca, we put patients first and strive to meet their unmet needs worldwide. Working here means being entrepreneurial, thinking big and working together to make the impossible a reality.</p>\n<p>Recognizing the importance of individualized flexibility, our ways of working allow employees to balance personal and work commitments while ensuring we continue to create a strong culture of collaboration and teamwork by engaging face-to-face in our offices 3 days a week.</p>\n<p>Our head office is purposely designed with collaboration in mind, providing space where teams can come together to strategize, brainstorm and connect on key projects.</p>\n<p>As a Senior Clinical Data Manager II, you will be responsible for coordination of the Clinical Data Management (CDM) deliverables on assigned clinical studies and may be an expert on CDM processes, standards, and technology.</p>\n<p>You will coordinate the Clinical Data Management deliverables on assigned studies depending on the relevant model and DM Vendor. Takes accountability and serves as the first line of contact at the study level.</p>\n<p>Demonstrates leadership and operational knowledge in the planning and delivery of CDM deliverables at a study level potentially under mentorship from a Project Data Manager.</p>\n<p>Communicates and collaborates effectively with all study team members. Primary point of contact for DM vendor and provides guidance and supervision to Lead Data Manager/DM Team Lead working on the study (CRO or in-house).</p>\n<p>Oversight of day-to-day operational aspects of CDM for assigned studies; Responsible to identify risks and collaborate with the DM Vendor to mitigate the risk. Escalates issues/risks when necessary.</p>\n<p>Understands corporate, therapeutic/indication or program specific data capture AZ standards.</p>\n<p>Provide input into CDM related activities associated with regulatory inspections/audits for assigned studies.</p>\n<p>Responsible for compliance to Trial Master File requirements relating to DM Vendor - Support Senior Leaders to oversee CDM Vendor performance, depending on relevant model. Review, assess and manage DM Vendor delivery against KPIs, budget and overall performance.</p>\n<p>Oversees vendor timelines and milestone deliverables for the assigned studies. Ensures DM Vendor billing is accurate and gives recommendations for payment of invoices.</p>\n<p>Drive adherence to AZ CDM standards and processes for data quality and consistency of data capturing for assigned studies.</p>\n<p>Demonstrates willingness to take on ad-hoc activities consistent with current CDM work experience.</p>\n<p>Ensures relevant training is completed prior to performing tasks.</p>\n<p>Mentoring junior Clinical Data Management colleagues - Performs CDM related ad-hoc requests from Line Manager.</p>\n<p>Essential Skills/Experience:</p>\n<p>Minimum of university or college degree in the life sciences or related subject, pharmacy, nursing or equivalent relevant degree</p>\n<p>Minimum of 5 years of Clinical Data Management and experience in the Biotech/Pharma/CRO industry</p>\n<p>Demonstrated current understanding of Good Clinical Data Management Practices (GCDMP) and relevant regulatory requirements</p>\n<p>Demonstrated experience of clinical databases, different clinical data management systems and electronic data capture (EDC)</p>\n<p>Demonstrate understanding and experience in query management process and reconciliation activities</p>\n<p>Ability to work flexibly on simultaneous projects and proactively manage time to meet own deadlines.</p>\n<p>Excellent written and verbal communication skills</p>\n<p>Ability to work in a global team environment</p>\n<p>Excellent organizational analytical skills and high attention to detail</p>\n<p>Desirable Skills/Experience:</p>\n<p>Demonstrated knowledge of clinical and pharmaceutical drug development process</p>\n<p>Demonstrated understanding of clinical data system design / development / validation and system interoperability.</p>\n<p>Demonstrated ability to work effectively with external partners</p>\n<p>Understanding of database structures, programming languages, data standards (CDISC) and practices as they apply to CRF design, database development, data handling and reporting</p>\n<p>Knowledge of SQL or SAS software</p>\n<p>Experience leading clinical studies as Data Management Lead</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_37654a6b-cb7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Clinical Data Management","sameAs":"https://astrazeneca.eightfold.ai","logo":"https://logos.yubhub.co/astrazeneca.eightfold.ai.png"},"x-apply-url":"https://astrazeneca.eightfold.ai/careers/job/563877689844672","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Clinical Data Management","Good Clinical Data Management Practices (GCDMP)","Clinical databases","Electronic data capture (EDC)","Query management process","Reconciliation activities","Global team environment","Organizational analytical skills","High attention to detail"],"x-skills-preferred":["Clinical and pharmaceutical drug development process","Clinical data system design / development / validation","System interoperability","Database structures","Programming languages","Data standards (CDISC)","CRF design","Database development","Data handling and reporting","SQL or SAS software"],"datePosted":"2026-04-18T22:12:55.410Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Durham, North Carolina, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"Clinical Data Management, Good Clinical Data Management Practices (GCDMP), Clinical databases, Electronic data capture (EDC), Query management process, Reconciliation activities, Global team environment, Organizational analytical skills, High attention to detail, Clinical and pharmaceutical drug development process, Clinical data system design / development / validation, System interoperability, Database structures, Programming languages, Data standards (CDISC), CRF design, Database development, Data handling and reporting, SQL or SAS software"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b6fdfd50-5de"},"title":"Threat Detection Engineer","description":"<p>Do you have a passion for hunting malicious activities in the background of business as usual and figuring out how to detect and respond to new threats?</p>\n<p>Millennium SOC is looking for an experienced Threat Detection Engineer to drive our best-in-class posture.</p>\n<p>This is a highly technical role, and successful candidates will have demonstrable knowledge and experience across a range of business and security technologies within a fast-paced organisation.</p>\n<p><strong>Principal Responsibilities</strong></p>\n<ul>\n<li>Identify modern evolving threats and develop new detection and response approaches</li>\n<li>Create and operate high-fidelity detections mechanisms that drive efficient, effective and repeatable response</li>\n<li>Own, operate and automate detection and response workflows, that enable the team to focus on strategic objectives</li>\n<li>Lead Information Security response activities for the firm</li>\n<li>Work across business and technology teams to deliver positive outcomes across the firm</li>\n<li>Explain complex technology and information security related concepts to a wide range of stakeholders</li>\n<li>Enforce security policies and procedures by administering and monitoring appropriate systems, events and answering stakeholder queries</li>\n<li>Monitor new and emerging security and privacy related technologies, trends, issues, and solutions and assess their applicability to Millennium key business initiatives and business strategies</li>\n<li>Ensure Millennium Information Security capabilities remain fit for purpose and evolve to meet the changing threat landscape</li>\n</ul>\n<p><strong>Qualifications/Skills Required</strong></p>\n<ul>\n<li>Bachelor or master’s degree in computer science or cyber security with strong IT background or equivalent demonstrable experience</li>\n<li>3 years’ experience working in a security engineering role, financial industry experience preferred</li>\n<li>Experience in creating detections in modern query languages (KQL, SQL, SPL)</li>\n<li>Possesses security certifications (Security+, OSCP, CISSP, CEH, GCIA, GCIH)</li>\n<li>Experience with modern security tooling across security domains; network, endpoint, data, identity and cloud</li>\n<li>Experience in standard enterprise technology stack, Active Directory, Entra, Group Policy, Intune, DNS, TCP/IP, PKI, Microsoft 365, Windows, Linux, MacOS, etc.</li>\n<li>Ability to handle sensitive and/or confidential materials with appropriate discretion</li>\n<li>Required scripting, development and automation skills using PowerShell or Python and proficient development tools</li>\n<li>Experience in OSINT, Threat hunting and analysing malicious emails</li>\n<li>Able to prioritise in a fast moving, high pressure, constantly changing environment</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b6fdfd50-5de","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT Infrastructure","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954323758","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Bachelor or master’s degree in computer science or cyber security","3 years’ experience working in a security engineering role","Experience in creating detections in modern query languages (KQL, SQL, SPL)","Possesses security certifications (Security+, OSCP, CISSP, CEH, GCIA, GCIH)","Experience with modern security tooling across security domains; network, endpoint, data, identity and cloud"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:52.812Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Finance","skills":"Bachelor or master’s degree in computer science or cyber security, 3 years’ experience working in a security engineering role, Experience in creating detections in modern query languages (KQL, SQL, SPL), Possesses security certifications (Security+, OSCP, CISSP, CEH, GCIA, GCIH), Experience with modern security tooling across security domains; network, endpoint, data, identity and cloud"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8610ea3d-93b"},"title":"Cloud Platform Engineer","description":"<p>The Business Development/Management Technology team at FIC &amp; Risk Technology is building and operating platforms that support recruiting, hiring, and onboarding of investment professionals. We are currently integrating multiple legacy and new systems into a unified, cloud-native platform to standardize processes, workflows, and data models across the organisation.</p>\n<p>This integration will enable seamless collaboration between teams and provide reliable, scalable data for analytics and reporting. We are looking for a Cloud Platform Engineer to design, build, and operate our AWS-based infrastructure and data platforms, using modern DevOps practices, infrastructure as code, and secure, well-engineered services in Python and C#.</p>\n<p>The successful candidate will collaborate with global technology and business teams to design cloud-native solutions that support business development and onboarding workflows. They will partner with global stakeholders to understand requirements and translate them into secure, scalable AWS architectures and platform capabilities.</p>\n<p>Key responsibilities include leading the end-to-end delivery of cloud and platform features, including design, implementation (Python/C#), infrastructure as code, testing, and deployment using DevOps practices.</p>\n<p>We are looking for a highly skilled engineer with 6+ years of experience in software or platform engineering, with significant time spent building and operating solutions in cloud environments (AWS preferred).</p>\n<p>The ideal candidate will have strong hands-on programming experience in Python and C#, with solid understanding of object-oriented design, design patterns, service-oriented / microservices architectures, concurrency, and SOLID principles.</p>\n<p>They will also have proven experience designing and operating AWS-based platforms (e.g., EC2, ECS/EKS, Lambda, S3, RDS, IAM) using infrastructure as code (Terraform, CloudFormation, or CDK).</p>\n<p>In addition, the successful candidate will have practical experience implementing DevOps practices and CI/CD pipelines (e.g., Jenkins, GitHub Actions, Azure DevOps), including automated testing, security scanning, and deployment.</p>\n<p>Experience supporting data science and analytics platforms, including orchestration tools such as Airflow, distributed processing engines such as Spark, and cloud-native data pipelines is also required.</p>\n<p>Good understanding of SQL and core database concepts; familiarity with AWS analytics services (e.g., Glue, EMR, Redshift, Athena) is a plus.</p>\n<p>Awareness of cloud security best practices, including IAM, network security, data encryption, and secure configuration management is also necessary.</p>\n<p>Strong problem-solving and analytical skills; demonstrated ability to take ownership, deliver in a fast-paced environment, and collaborate effectively with global teams is essential.</p>\n<p>Excellent communication skills, with ability to work closely with both technical and non-technical stakeholders is also required.</p>\n<p>Experience estimating, monitoring, and optimizing AWS infrastructure costs, including use of tools such as AWS Cost Explorer, AWS Budgets, and cost-allocation tagging strategies is desirable.</p>\n<p>Experience designing and operating workloads across multiple cloud environments and on-premises, using centralized policies, governance, and controls to support business-aligned teams is also beneficial.</p>\n<p>Working knowledge of networking across on-premises and cloud environments, including VPC design, subnets, routing, VPNs/Direct Connect, load balancing, DNS, and network security controls is necessary.</p>\n<p>Nice to have experience with additional big data tools or platforms (e.g., Kafka, Databricks, Snowflake, Flink).</p>\n<p>Familiarity with Capital Markets concepts and operating models is also beneficial.</p>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p>When finalising an offer, we take into consideration an individual&#39;s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8610ea3d-93b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955139979","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["AWS","Python","C#","DevOps","Infrastructure as Code","Cloud Security","SQL","Database Concepts","Networking"],"x-skills-preferred":["Airflow","Spark","Kafka","Databricks","Snowflake","Flink","Capital Markets"],"datePosted":"2026-04-18T22:12:50.548Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"AWS, Python, C#, DevOps, Infrastructure as Code, Cloud Security, SQL, Database Concepts, Networking, Airflow, Spark, Kafka, Databricks, Snowflake, Flink, Capital Markets","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3a94ff8f-44b"},"title":"Treasury Operations Specialist","description":"<p>We are seeking a driven and detail-oriented Operations Specialist to partner with our Treasury Technology team to design and build out our proprietary cash management &amp; margin platform, develop data automation solutions, and drive process re-engineering efforts.</p>\n<p>As a key member of the Treasury Operations team, you will play a crucial role in system design &amp; implementation of our proprietary cash management, margin &amp; collateral management platform. This includes collaborating within the Treasury Operations team to define &amp; document system requirements, performing business process analysis, data analysis, UAT testing, data migrations &amp; reconciliations, producing user interface mockups, and data and process flow diagrams.</p>\n<p>You will also develop data automation solutions, reports, and reconciliations using Alteryx and SQL to support the Treasury Operations team. Additionally, you will coordinate and collaborate closely with internal teams across Operations, Middle Office, Information Technology, Legal, Compliance, and Finance, as well as external parties, to deliver projects and solutions.</p>\n<p>Your responsibilities will also include driving process re-engineering, including documentation, analysis, and presentation of current-state process flows and recommended target state flows. You will perform data analysis to provide business transparency, identify deficiencies and support platform implementations &amp; enhancements.</p>\n<p>To excel in this role, you should possess strong technical skills in data management, coupled with a solid understanding of business processes and workflows. Effective communication, project management, and collaboration skills are also essential.</p>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package. The estimated base salary range for this position is $70,000 to $160,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3a94ff8f-44b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Core Operations","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954693925","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$70,000 to $160,000","x-skills-required":["data management","business processes and workflows","Alteryx","SQL","project management","collaboration"],"x-skills-preferred":["Treasury systems","cash management","SWIFT payment formatting","margin"],"datePosted":"2026-04-18T22:12:48.161Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"data management, business processes and workflows, Alteryx, SQL, project management, collaboration, Treasury systems, cash management, SWIFT payment formatting, margin","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":160000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9ca997fb-218"},"title":"Quantitative Developer","description":"<p>We are building a world-class systematic data platform that will power the next generation of our systematic portfolio engines.</p>\n<p>The systematic data group is looking for a Quantitative Developer to join our growing team. The team consists of content specialists, data scientists, engineers, and quant developers who are responsible for discovering, maintaining, and analysing sources of alpha for our portfolio managers.</p>\n<p>The role builds on individual&#39;s knowledge and skills in four key areas of quantitative investing: data, statistics, technology, and financial markets.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Use finance knowledge and statistical knowledge to analyse potential alpha sources and present findings to portfolio managers and quantitative analysts.</li>\n<li>Build quant tools to help portfolio managers research, evaluate, combine alphas, and understand risks.</li>\n<li>Design and maintain tools to evaluate and monitor data quality and integrity for a wide variety of data sources.</li>\n<li>Engage with vendors, brokers, and perform analytics to understand characteristics of datasets.</li>\n<li>Interact with portfolio managers and quantitative analysts to understand their use cases and recommend datasets to help maximise their profitability.</li>\n</ul>\n<p>Skills Required:</p>\n<ul>\n<li>3+ years of work experience as a financial engineer, data scientist, or quant developer.</li>\n<li>Strong knowledge of Python and/or C++, Java, C#.</li>\n<li>Familiarity with data pipeline engineering, ETL for large datasets, and scheduling tools like Airflow.</li>\n<li>Strong SQL and database experience including PL-SQL or T-SQL.</li>\n<li>Understanding of typical software development lifecycle and familiarity with: Linux, GitHub, CI/CD.</li>\n<li>Ph.D. or Masters in computer science, mathematics, statistics, or other field requiring quantitative analysis.</li>\n</ul>\n<p>Beneficial Skills and Experience:</p>\n<ul>\n<li>Understanding of risk models and performance attribution.</li>\n<li>Experience with financial markets such as equities and futures.</li>\n<li>Knowledge of statistical techniques and their usage.</li>\n</ul>\n<p>The estimated base salary range for this position is $165,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9ca997fb-218","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755952876477","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$165,000 to $250,000","x-skills-required":["Python","C++","Java","C#","data pipeline engineering","ETL","Airflow","SQL","database","Linux","GitHub","CI/CD","Ph.D.","Masters"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:44.538Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, C++, Java, C#, data pipeline engineering, ETL, Airflow, SQL, database, Linux, GitHub, CI/CD, Ph.D., Masters","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4db63d33-c03"},"title":"Senior Technical Business Analyst - Algorithmic Execution","description":"<p>We are seeking a Senior Technical Business Analyst to join our team in New York. As a key member of our Execution Services and Central Liquidity Strategies teams, you will work closely with Portfolio Managers to build next-generation algorithmic solutions.</p>\n<p>In this role, you will play a critical role at the intersection of technology and trading, delivering order routing and internal liquidity products that optimize execution and drive savings for the firm.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Work with Execution Services and other internal stakeholders to gather and synthesise requirements across our algorithmic trading strategies.</li>\n</ul>\n<ul>\n<li>Drive the strategic expansion of the internal algo platform into new markets and asset classes.</li>\n</ul>\n<ul>\n<li>Analyse upstream/downstream data dependencies and create the necessary documented requirements, JIRAs, and project plans for core execution components.</li>\n</ul>\n<ul>\n<li>Create user guides and technical documentation to support the onboarding of new PMs and desks across our platforms.</li>\n</ul>\n<ul>\n<li>Create and maintain product roadmaps and artefacts required to manage stakeholder expectations across Trading and Technology.</li>\n</ul>\n<ul>\n<li>Manage day-to-day project deliverables; highlight, escalate, and resolve issues, conflicts, and roadblocks in a fast-paced trading environment.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>10 years of experience as a Technical Business Analyst or Project Manager in an enterprise-level FinTech environment.</li>\n</ul>\n<ul>\n<li>5+ years of relevant trading technology experience, ideally on the buyside, and comfortable interacting with front-office, non-technical personnel.</li>\n</ul>\n<ul>\n<li>Subject matter expertise in Electronic Execution and Market Microstructure (Equities required, Futures highly preferred).</li>\n</ul>\n<ul>\n<li>Impactful individual contributor: Be able to lead a wide range of projects front to back.</li>\n</ul>\n<ul>\n<li>Technical skills: Self-sufficient with SQL for trade data analysis and troubleshooting. Experience with KDB/Q and/or Python for data analysis preferred. Experience with sequencer-based platforms is also desired.</li>\n</ul>\n<ul>\n<li>Communication: Strong communication skills and the ability to work effectively in a team environment.</li>\n</ul>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4db63d33-c03","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755938267676","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["SQL","KDB/Q","Python","Electronic Execution","Market Microstructure"],"x-skills-preferred":["Sequencer-based platforms"],"datePosted":"2026-04-18T22:12:44.525Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Finance","skills":"SQL, KDB/Q, Python, Electronic Execution, Market Microstructure, Sequencer-based platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8b09f8c4-a35"},"title":"Salesforce Team Lead","description":"<p>We are seeking a Salesforce Team Lead to implement Salesforce for Millennium, focusing on Business Development and Business Management functions. These functions support the recruitment and support of Portfolio Managers. The ideal candidate will have strong technical skills in Salesforce, as well as a credible level of functional depth.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Develop and operate a recruiting platform for portfolio managers.</li>\n<li>Participate in the design of various functions for Salesforce CRM solution.</li>\n<li>Operate and further develop existing CRM and existing supporting products and applications.</li>\n<li>Develop interfaces between CRM software and other systems internal and external to Millennium.</li>\n<li>Lead the environment strategy as well as deployment strategy for the CRM solution.</li>\n<li>Stay up to date with the latest Salesforce releases, features, and best practices.</li>\n</ul>\n<p>Qualifications/Skills Required:</p>\n<ul>\n<li>10+ years of experience in Salesforce sales cloud development/support.</li>\n<li>4-7 years software development experience.</li>\n<li>3+ years of C#/.NET with proficiency in a web framework such as ASP.NET MVC.</li>\n<li>2+ years of WebUI/JavaScript/HTML/CSS with proficiency in at least one web framework such as Angular, React/Redux, or Ember.</li>\n<li>In-depth knowledge of Salesforce Feedback Management, Salesforce Digital Experiences (Communities) and best practices.</li>\n<li>Drive AI Innovation: Proactively research, propose, and prototype AI-driven solutions that enhance business processes, customer experience, or operational efficiency within Salesforce.</li>\n<li>Track record of proposing or leading AI initiatives that resulted in measurable business impact.</li>\n<li>Experience building and consuming RESTful services.</li>\n<li>Experience building integration with third-party systems to a Salesforce CRM.</li>\n<li>Able to lead the Salesforce deployments.</li>\n<li>Design, customize and manage security entitlements for Salesforce sales cloud.</li>\n<li>Development experience in Microsoft SQL server, building complex SQL and stored procedures.</li>\n<li>Strong and effective interpersonal skills with proven ability to develop positive relationships with business partners.</li>\n<li>Highly analytical with good problem solving skills; able to work independently in a fast-paced environment.</li>\n</ul>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8b09f8c4-a35","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954905556","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Salesforce sales cloud development/support","Software development experience","C#/.NET","WebUI/JavaScript/HTML/CSS","Salesforce Feedback Management","Salesforce Digital Experiences (Communities)","RESTful services","Integration with third-party systems","Microsoft SQL server","Complex SQL and stored procedures"],"x-skills-preferred":["Integration experience with real time messaging platforms like Kafka","Experience working with front office business data and processes","Exposure to Agent force"],"datePosted":"2026-04-18T22:12:36.839Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Technology","skills":"Salesforce sales cloud development/support, Software development experience, C#/.NET, WebUI/JavaScript/HTML/CSS, Salesforce Feedback Management, Salesforce Digital Experiences (Communities), RESTful services, Integration with third-party systems, Microsoft SQL server, Complex SQL and stored procedures, Integration experience with real time messaging platforms like Kafka, Experience working with front office business data and processes, Exposure to Agent force","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_528bf454-d13"},"title":"Data Analytics Engineer","description":"<p>We are seeking a Senior Analytics Engineer to join our team. As a key member of our data organization, you will be responsible for transforming raw data into a strategic asset by designing high-performance data models that power our financial reporting, product forecasting, and GTM strategy.</p>\n<p>Your 12-Month Journey</p>\n<p>During the first 3 months, you will learn about our existing stack (GCP, BigQuery, Airbyte, dbt), core business data models, and understand the current pain points in our data flow. You will deliver and optimize your first high-priority models for product usage and financial reporting. You will partner with the Data Engineer to align on the new infrastructure roadmap.</p>\n<p>Within 6 months, you will implement a robust semantic layer to standardize KPIs across the company and enable AI-readiness and advanced natural language querying.</p>\n<p>After 1 year, you will fully own the company&#39;s data modeling architecture, ensuring it is prepared for AI and machine learning applications. You will act as a strategic advisor to department heads, using data to help shape the company&#39;s long-term growth and forecasting strategies.</p>\n<p>What You&#39;ll Be Doing</p>\n<p>Strategic Data Product Ownership: Manage the end-to-end lifecycle of our internal data products. You will partner with stakeholders to translate complex business questions into technical requirements, selecting the right tools to ensure our reporting is scalable, accessible, and high-impact.</p>\n<p>Advanced Analytics Engineering: Design, build, and maintain our core data models using dbt Labs. You will own the logic for mission-critical datasets, including financial reporting, churn forecasting, and reverse-ETL flows that sync warehouse data back into our business tools (e.g., Planhat, HubSpot).</p>\n<p>Data Governance &amp; Semantic Layering: Act as the guardian of &#39;The Truth.&#39; You will implement data governance standards and build our semantic layer to ensure metrics are consistent across the company.</p>\n<p>Data Democratization &amp; Enablement: In collaboration with RevOps, you will design and deliver training programs and documentation. Your goal is to empower users across Finance, Product, and GTM to independently navigate data products and derive their own insights.</p>\n<p>Collaboration: You will be the central hub of our data organization. You will work daily with the Data Engineer to align on the roadmap, while frequently consulting with Finance, GTM, and Product leaders to ensure our data products solve their most pressing problems.</p>\n<p>What You Bring</p>\n<p>Solid experience in Analytics Engineering, Data Analysis, or Data Engineering, with a track record of independently delivering data products that enable reporting, decision-making, and CDP use cases.</p>\n<p>You are an expert in SQL and understand how to write performant, modular code. Familiarity with Python and Git for optimizing and versioning data transformations is a significant advantage.</p>\n<p>Deep, hands-on experience with dbt and BigQuery is a must. You should also be comfortable navigating ELT tools like Airbyte or Fivetran.</p>\n<p>Commercially savvy: you understand the business. You can spot opportunities where data can improve ARR, reduce churn, or optimize spend.</p>\n<p>You thrive in fast-paced environments and are comfortable creating structure out of the uncertainty of a scaling company.</p>\n<p>Strong project management and stakeholder management skills. You are a &#39;bilingual&#39; communicator who can discuss warehouse schemas with an engineer and ARR growth with a CFO.</p>\n<p>Fluency in English, both written and spoken, at a minimum C1 level</p>\n<p>What We Offer</p>\n<p>Flexibility to work from home in the Netherlands and from our beautiful canal-side office in Amsterdam</p>\n<p>A chance to be part of and shape one of the most ambitious scale-ups in Europe</p>\n<p>Work in a diverse and multicultural team</p>\n<p>€1,500 annual training budget plus internal training</p>\n<p>Pension plan, travel reimbursement, and wellness perks</p>\n<p>28 paid holiday days + 2 additional days to relax in 2026</p>\n<p>Work from anywhere for 4 weeks/year</p>\n<p>An inclusive and international work environment with a whole lot of fun thrown in!</p>\n<p>Apple MacBook and tools</p>\n<p>€200 Home Office budget</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_528bf454-d13","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Tellent","sameAs":"https://careers.tellent.com","logo":"https://logos.yubhub.co/careers.tellent.com.png"},"x-apply-url":"https://careers.tellent.com/o/data-analytics-engineer","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"EUR 70000–90000 / year","x-skills-required":["SQL","dbt","BigQuery","Airbyte","Python","Git","ELT tools","Data governance","Semantic layering","Data democratization","Enablement"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:13.210Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Amsterdam"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, dbt, BigQuery, Airbyte, Python, Git, ELT tools, Data governance, Semantic layering, Data democratization, Enablement","baseSalary":{"@type":"MonetaryAmount","currency":"EUR","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":90000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_21f5f6c3-734"},"title":"Data Engineer","description":"<p>About the Role We are at a pivotal scaling point where our data ambitions have outpaced our current setup, and we need a Data Engineer to architect the professional-grade foundations of our platform.</p>\n<p>This role exists to bridge the gap between &quot;getting data&quot; and &quot;engineering data,&quot; moving us from manual syncs to a fully automated ecosystem. By building custom pipelines and implementing a robust orchestration layer, you will directly enable our Operations teams and leadership to transition from basic reporting to sophisticated, AI-ready data products.</p>\n<p>Your primary focus will be on Infrastructure-as-Code, orchestration, and building a resilient &quot;plumbing&quot; system that serves as the backbone for our entire Product and GTM strategy.</p>\n<p>Your 12-Month Journey During the first 3 months: you will learn about our existing stack (GCP, BigQuery, Airbyte, dbt) and understand the current pain points in our data flow. You will identify and execute &quot;low-hanging fruit&quot; improvements to our product usage analytics, providing immediate value to the Product and GTM teams. You’ll begin designing the blueprint for our custom data pipelines and the migration strategy for moving our infrastructure into Terraform.</p>\n<p>Within 6 months: You will have deployed our new orchestration layer (e.g., Airflow or Dagster) and successfully transitioned our first set of custom pipelines to production. Collaborating with the Analytics Engineer, you will enable a unified view of our customer journey by successfully merging product usage data with CRM and billing data. At this point, a significant portion of our data infrastructure will be defined as code, reducing manual overhead and increasing deployment reliability.</p>\n<p>After 1 year: you will take full strategic ownership of the data platform and its long-term architecture. You will act as the go-to technical expert for the leadership team, advising on the scalability of new data-driven features. You will lay the groundwork for AI and Machine Learning initiatives by ensuring our data warehouse has the right quality controls, governance, and low-latency access patterns in place.</p>\n<p>What You’ll Be Doing Architect Scalable Infrastructure-as-Code: Take our existing foundations to the next level by migrating all GCP and BigQuery resources into Terraform. You will establish automated CI/CD patterns to ensure our entire data environment is reproducible, version-controlled, and enterprise-ready.</p>\n<p>Deploy State-of-the-Art Pipelines: Design, deploy, and operate high-quality production ELT pipelines. You will implement a modern orchestration layer (e.g., Airflow or Dagster) to build custom Python-based integrations while maintaining and optimizing our existing syncs.</p>\n<p>Champion Data Quality &amp; Performance: Act as the guardian of our data platform. You will implement rigorous testing and monitoring protocols to ensure data is accurate and timely. You will proactively identify BigQuery bottlenecks, optimizing query performance and resource utilization.</p>\n<p>Technical Roadmap &amp; Ownership: scope and architect end-to-end data flows from production source to warehouse. Manage your own technical backlog, prioritizing infrastructure stability over technical debt. You will ensure platform security and SOC2 compliance through PII masking, data contracts, and robust access controls.</p>\n<p>Collaboration: You will work in a tight loop with the Analytics Engineer to turn raw data into actionable products. You will partner daily with DataOps and RevOps to understand business requirements, with occasional strategic syncs with DevOps and R&amp;D to align on production schema changes and global infrastructure standards.</p>\n<p>What You Bring Solid experience in Data Engineering, with a track record of building and evolving data ingestion infrastructure in cloud environments. The Modern Data Stack: Familiarity with dbt and Airbyte/Fivetran. You understand how these tools fit into a broader ecosystem. Expertise in BigQuery (partitioning, clustering, IAM) and the broader GCP ecosystem; Infrastructure-as-Code (Terraform). Hands-on experience with Airflow, Dagster, or similar orchestration tools. You know how to design DAGs that are resilient and easy to debug. DevOps practices in the data context: familiarity with CI/CD best practices as they apply to data (data testing, automated deployments). Programming: Expert-level Python and advanced SQL. You are comfortable writing clean, testable, and modular code. Comfortable in a fast-paced environment Project management skills: capable of managing stakeholders, explaining complicated technical trade-offs to non-technical users, and taking care of own project scoping and backlog management. Fluency in English, both written and spoken, at a minimum C1 level</p>\n<p>What We Offer Flexibility to work from home in the Netherlands and from our beautiful canal-side office in Amsterdam A chance to be part of and shape one of the most ambitious scale-ups in Europe Work in a diverse and multicultural team €1,500 annual training budget plus internal training Pension plan, travel reimbursement, and wellness perks 28 paid holiday days + 2 additional days to relax in 2026 Work from anywhere for 4 weeks/year An inclusive and international work environment with a whole lot of fun thrown in! Apple MacBook and tools €200 Home Office budget</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_21f5f6c3-734","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Tellent","sameAs":"https://careers.tellent.com","logo":"https://logos.yubhub.co/careers.tellent.com.png"},"x-apply-url":"https://careers.tellent.com/o/data-engineer","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"EUR 70000–90000 / year","x-skills-required":["Data Engineering","Cloud environments","dbt","Airbyte/Fivetran","BigQuery","GCP ecosystem","Infrastructure-as-Code","Terraform","Airflow","Dagster","Python","SQL","CI/CD best practices","DevOps practices"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:06.548Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Amsterdam"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Data Engineering, Cloud environments, dbt, Airbyte/Fivetran, BigQuery, GCP ecosystem, Infrastructure-as-Code, Terraform, Airflow, Dagster, Python, SQL, CI/CD best practices, DevOps practices","baseSalary":{"@type":"MonetaryAmount","currency":"EUR","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":90000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e8aabc91-c80"},"title":"Assistant Manager of Data Analytics","description":"<p>We are seeking an experienced professional to join our team in Shanghai. As Assistant Manager of Data Analytics, you will focus on using data and analytics to drive business activities and outcomes that improve or transform customer strategy, customer segmentation, predictive models, and marketing campaigns.</p>\n<p>Principal Responsibilities: The role holder will conduct customer strategy analysis focusing on acquisition, activation, retention, conversion, and LTV, and deliver actionable insights. Build and maintain customer segmentation frameworks to support targeted and personalized marketing and operations. Leverage advanced data analytics tools and methodologies to develop, validate, and optimize predictive models, contributing to generate high-quality leads. Analyze customer journey, conversion funnels, and drop-off points to identify bottlenecks and recommend experience improvements. Evaluate the performance of marketing campaigns, membership programs, loyalty initiatives, and promotional strategies by measuring ROI, conversion rate, and engagement metrics. Partner with product, marketing, operations, and customer teams to translate data insights into executable strategies and drive business decisions. Support the business team&#39;s campaign needs, including RM lead generation and manual SMS outreach. Develop and maintain customer-focused dashboards, KPIs, and reporting systems.</p>\n<p>To be successful in the role, you should meet the following requirements: Minimum of 5 years&#39; experience in one or multiple skills in data/business analytics in the financial or digital domains. Demonstrated experience in process and analysis of large amounts of data using one of these: Python, R, SQL, or SAS; on environments such as AWS, Google Cloud, or Hadoop. Knowledge and experience in AI, big data, machine learning, or predictive algorithms, statistics modeling, and data mining. Excellent communication and teamwork skills, able to collaborate effectively with different departments and stakeholders. Strong problem-solving skills and innovative thinking, able to translate complex business problems into data analytics solutions. Proven experience in one or more of: customer segmentation, digital marketing, data science, portfolio analytics, use of open-source data in analyses. Good English communication skills, able to collaborate effectively with domestic and international teams.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e8aabc91-c80","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC International Wealth and Premier Banking","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610677890","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","R","SQL","SAS","AWS","Google Cloud","Hadoop","AI","big data","machine learning","predictive algorithms","statistics modeling","data mining"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:11:33.642Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Shanghai"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"Python, R, SQL, SAS, AWS, Google Cloud, Hadoop, AI, big data, machine learning, predictive algorithms, statistics modeling, data mining"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_083d8199-705"},"title":"Iseries/Consultant Specialist","description":"<p>Join HSBC and fulfil your potential in a career that shines brighter than others.</p>\n<p>Whether you want a career that could take you to the top, or simply take you in an exciting new direction, HSBC offers opportunities, support and rewards that will take you further.</p>\n<p>In this role, you will be an approachable and supportive team member with a collaborative attitude within a demanding, maturing Agile environment.</p>\n<p>You will have great communication skills, conveying your thoughts, ideas, and opinions clearly and concisely face-to-face or virtually to all levels up and down stream.</p>\n<p>Regularly, you will demonstrate qualities such as drive, motivation, determination, dedication, resiliency, honesty, and enthusiasm.</p>\n<p>You will be culturally aware and sensitive, and be flexible under pressure.</p>\n<p>Strong analytical and problem-solving skills are required, along with excellent verbal and written communication skills.</p>\n<p>To be successful in this role, you will need:</p>\n<ul>\n<li>8+ years of hands-on expertise in RPG400/IV, RPG Free, CL, SQL, Embedded SQL, Query &amp; ILE</li>\n</ul>\n<ul>\n<li>Recent and advanced experience with RPG (ILE/FREE) using Procedures, Service Programs, and Functions</li>\n</ul>\n<ul>\n<li>A strong understanding and working experience with CI/CD and available tools, such as Jenkins, Sonar, RDI</li>\n</ul>\n<ul>\n<li>Working experience in an agile environment</li>\n</ul>\n<ul>\n<li>The ability to quickly acquire new skills and tools</li>\n</ul>\n<ul>\n<li>The ability to resolve critical issues in a timely manner</li>\n</ul>\n<ul>\n<li>Experience releasing to production what you developed and then supporting it</li>\n</ul>\n<ul>\n<li>Hands-on experience of Application Design &amp; Development in BFSI/Banking domain is desirable</li>\n</ul>\n<ul>\n<li>Change management experience and familiarity with a change management tool</li>\n</ul>\n<ul>\n<li>Understanding of DevOps concepts and techniques</li>\n</ul>\n<p>You’ll achieve more when you join HSBC.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_083d8199-705","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610663254","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["RPG400/IV","RPG Free","CL","SQL","Embedded SQL","Query & ILE","Jenkins","Sonar","RDI","Agile"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:11:26.497Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"RPG400/IV, RPG Free, CL, SQL, Embedded SQL, Query & ILE, Jenkins, Sonar, RDI, Agile"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f7aeee90-9b7"},"title":"Technical Specialist (Java, Microservices) / Associate Director, Software Engineering","description":"<p>Join HSBC and help you stand out in your career. We offer opportunities, support and rewards that will take you further. As an Associate Director, Software Engineering, you will lead the development and implementation of Microservices-based solutions using Java. You will also architect and design scalable, distributed systems with high availability, collaborate with cross-functional teams to gather requirements and deliver solutions, ensure code quality through best practices, code reviews, and automated testing, mentor and guide team members in technical aspects and career growth, troubleshoot and resolve complex technical issues in production environments, stay updated with emerging technologies and recommend their adoption, navigate a dynamic ecosystem to deliver change effectively, demonstrating initiative, self-motivation, and drive, and exhibit tenacity and determination to clarify business requirements and deliver solutions in occasionally challenging circumstances.</p>\n<p>To be successful in this role, you should have strong proficiency in Java (Java 21 preferred), hands-on experience with Microservices architecture and frameworks (e.g., Spring Boot, Spring Cloud), expertise in RESTful APIs, messaging systems (e.g., Kafka, Hazelcast), and containerization (e.g., Docker, Kubernetes), solid understanding of cloud platforms (e.g., Kubernetes platform, GCP and AWS), hands-on experience with CI/CD pipelines and DevOps practices, knowledge of database technologies (SQL and NoSQL), payment&#39;s domain experience and clearing scheme experience, excellent problem-solving and communication skills, hands-on experience in both SDLC and Agile methodologies, familiarity with monitoring tools (e.g., Prometheus, Grafana, Splunk), and certifications in Java or cloud technologies are a plus.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f7aeee90-9b7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610662228","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Microservices architecture","Spring Boot","Spring Cloud","RESTful APIs","Kafka","Hazelcast","Docker","Kubernetes","CI/CD pipelines","DevOps practices","database technologies","SQL","NoSQL","payment's domain experience","clearing scheme experience"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:10:55.397Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hyderabad, Telangana, India · Bangalore, Karnataka, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Microservices architecture, Spring Boot, Spring Cloud, RESTful APIs, Kafka, Hazelcast, Docker, Kubernetes, CI/CD pipelines, DevOps practices, database technologies, SQL, NoSQL, payment's domain experience, clearing scheme experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_aee9464f-897"},"title":"Technical Specialist (Java, Microservices) / Associate Director, Software Engineering","description":"<p>We are currently seeking an experienced professional to join our team in the role of a Associate Director, Software Engineering.</p>\n<p>In this role, you will lead the development and implementation of Microservices-based solutions using Java. You will also architect and design scalable, distributed systems with high availability, collaborate with cross-functional teams to gather requirements and deliver solutions, ensure code quality through best practices, code reviews, and automated testing, mentor and guide team members in technical aspects and career growth, troubleshoot and resolve complex technical issues in production environments, stay updated with emerging technologies and recommend their adoption, navigate a dynamic ecosystem to deliver change effectively, demonstrating initiative, self-motivation, and drive, exhibit tenacity and determination to clarify business requirements and deliver solutions in occasionally challenging circumstances.</p>\n<p>To be successful in this role, you should meet the following requirements:</p>\n<ul>\n<li>Strong proficiency in Java (Java 21 preferred).</li>\n<li>Hands-on experience with Microservices architecture and frameworks (e.g., Spring Boot, Spring Cloud).</li>\n<li>Expertise in RESTful APIs, messaging systems (e.g., Kafka, Hazelcast), and containerization (e.g., Docker, Kubernetes).</li>\n<li>Solid understanding of cloud platforms (e.g., Kubernetes platform, GCP and AWS).</li>\n<li>Hands-on experience with CI/CD pipelines and DevOps practices.</li>\n<li>Knowledge of database technologies (SQL and NoSQL).</li>\n<li>Payment&#39;s domain experience and clearing scheme experience.</li>\n<li>Excellent problem-solving and communication skills.</li>\n<li>Hands-on experience in both SDLC and Agile methodologies.</li>\n<li>Familiarity with monitoring tools (e.g., Prometheus, Grafana, Splunk).</li>\n<li>Certifications in Java or cloud technologies are a plus.</li>\n</ul>\n<p>You&#39;ll achieve more when you join HSBC.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_aee9464f-897","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610662222","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Microservices","Spring Boot","Spring Cloud","RESTful APIs","Kafka","Hazelcast","Docker","Kubernetes","CI/CD pipelines","DevOps practices","database technologies","SQL","NoSQL","payment's domain experience","clearing scheme experience"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:10:31.133Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, Hyderabad"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Microservices, Spring Boot, Spring Cloud, RESTful APIs, Kafka, Hazelcast, Docker, Kubernetes, CI/CD pipelines, DevOps practices, database technologies, SQL, NoSQL, payment's domain experience, clearing scheme experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6bfc6b4-74f"},"title":"Senior Data Scientist - Marketing (all genders)","description":"<p>Join our Business Intelligence Department, a multidisciplinary group of Data Scientists, Analysts, and Data Engineers. Together, we build machine learning and analytics products that directly influence GMV, conversion, and retention.</p>\n<p>Within the department, we’re building a new Marketing Analytics team and are looking for a Senior Data Scientist to drive its data science initiatives. In this role, you’ll work closely with Analysts, Engineers, and Marketing stakeholders to develop and productionize advanced machine learning, statistical, and predictive models that improve marketing performance and drive measurable company growth.</p>\n<p>As a Senior Data Scientist – Marketing, you’ll take strong ownership of data science initiatives that directly shape our marketing strategy and growth. You will:</p>\n<p>Partner closely with Marketing, Marketing Analytics, and Marketing Technology to identify opportunities and translate business questions into scalable data science solutions.</p>\n<p>Lead the development of high-impact machine learning and statistical models for marketing use cases such as channel allocation, ad bidding, churn prediction, lifetime value, revenue attribution, and business metrics forecasting.</p>\n<p>Work end-to-end - from translating business questions into hypotheses to researching, building, validating, and deploying models.</p>\n<p>Run experiments and iterate in production: design A/B tests, monitor model performance, and continuously improve based on measured impact.</p>\n<p>Advance our MLOps practices with CI/CD pipelines, retraining workflows, lineage tracking, and documentation.</p>\n<p>Help define the team&#39;s roadmap and ways of working as a founding member of Marketing Analytics - your input will help shape this function.</p>\n<p>Act as a senior role model in the team, sharing best practices and helping raise the bar for data science at Holidu.</p>\n<p>We&#39;re looking for someone with 5+ years of experience as a Data Scientist, with clear ownership of projects that delivered measurable business impact. You should have a degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field, and strong expertise in machine learning, statistics, and predictive analytics, with hands-on experience using Python and SQL.</p>\n<p>Experience with marketing data science use cases such as attribution modeling, customer lifetime value prediction, churn modeling, or bid optimization is also required. You should have a solid understanding of marketing concepts across channels (e.g. Performance Marketing, SEO, CRM, Affiliate) and how data science can improve them.</p>\n<p>Additionally, you should have experience working with modern data stacks, ideally including AWS (Redshift, Athena, S3), Airflow, dbt, and Git. A collaborative mindset paired with great communication skills is essential, as you&#39;ll need to work with diverse stakeholders and explain complex topics in a simple way.</p>\n<p>AI proficiency is also a plus, as you&#39;ll be comfortable using AI to enhance coding, planning, and monitoring, and successfully integrating AI tools (such as Claude code, Codex, Copilot, etc.) into your workflow and teaching others to use them efficiently.</p>\n<p>If you&#39;re excited about the opportunity to shape the future of travel with products used by millions of guests and thousands of hosts, apply now!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6bfc6b4-74f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2510157","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Machine Learning","Statistics","Predictive Analytics","Python","SQL","Marketing Data Science","Attribution Modeling","Customer Lifetime Value Prediction","Churn Modeling","Bid Optimization"],"x-skills-preferred":["AI","CI/CD Pipelines","Retraining Workflows","Lineage Tracking","Documentation","Airflow","dbt","Git"],"datePosted":"2026-04-18T22:10:24.739Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, Statistics, Predictive Analytics, Python, SQL, Marketing Data Science, Attribution Modeling, Customer Lifetime Value Prediction, Churn Modeling, Bid Optimization, AI, CI/CD Pipelines, Retraining Workflows, Lineage Tracking, Documentation, Airflow, dbt, Git"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_02c944ab-f9e"},"title":"Senior Data Scientist - Dynamic Pricing & Revenue Management (all genders)","description":"<p>You&#39;ll be part of our new Dynamic Pricing &amp; Revenue Management team, working alongside a Data Analyst and Data Engineer. Together, you will work towards one core goal: helping hosts improve occupancy and earnings through a smart, dynamic and data driven pricing strategy.</p>\n<p>You&#39;ll work with a large and rich dataset, modern tooling, and teammates who care deeply about impact, collaboration, and learning together. This role is based in Munich with 3 office days per week.</p>\n<p>As a Senior Data Scientist, you&#39;ll take ownership of complex pricing and forecasting models and help us turn analytical ideas into real-world impact for hosts and Holidu. You will:</p>\n<ul>\n<li>Translate business questions into scientific, testable models and clear recommendations.</li>\n<li>Design, build and own machine learning, forecasting and predictive models for revenue management topics such as demand forecasting, price sensitivity, and conversion probability.</li>\n<li>Explore and develop dynamic pricing strategies (e.g. weekend pricing, early discounts, regional similarities) using data and experimentation.</li>\n<li>Collaborate closely with Data Analysts and Data Engineers to define datasets, features, and model requirements.</li>\n<li>Drive discussions around model choice, assumptions, and trade-offs, always keeping business impact in mind.</li>\n<li>Monitor model performance, iterate on results, and continuously improve accuracy and relevance.</li>\n<li>Act as a senior sparring partner in the team, sharing knowledge and raising the bar for data science practices.</li>\n</ul>\n<p>You&#39;ll have 5+ years of experience as a Data Scientist, solving a variety of different business problems. You&#39;ll have a strong background in statistics, forecasting, and machine learning. You&#39;ll be hands-on with Python and SQL, and confident working with large datasets. You&#39;ll have a strong interest in pricing, revenue optimization, or marketplace dynamics (prior revenue management experience is a plus, not a must).</p>\n<p>You&#39;ll be a self-starter: proactive, hungry to learn, and eager to make an impact. You&#39;ll be able to communicate complex ideas clearly and collaborate with technical and non-technical partners.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_02c944ab-f9e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2518625","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","SQL","Machine Learning","Forecasting","Predictive Modeling","Data Science","Data Analysis","Data Engineering"],"x-skills-preferred":["Dynamic Pricing","Revenue Optimization","Marketplace Dynamics","Cloud Computing","Big Data","Data Visualization"],"datePosted":"2026-04-18T22:10:08.998Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Machine Learning, Forecasting, Predictive Modeling, Data Science, Data Analysis, Data Engineering, Dynamic Pricing, Revenue Optimization, Marketplace Dynamics, Cloud Computing, Big Data, Data Visualization"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_52ea5e8c-da4"},"title":"Corporate Sales Associate","description":"<p>In compliance with applicable laws, HSBC is committed to employing only those who are authorised to work in the US. As a Corporate Sales Associate, you will support the Corporate Sales Chief Operating Officer (COO) function and Corporate Sales Team globally in Business Development, Sales Support, Client Service, Booking and execution, Market Research and Insights, Automation etc.</p>\n<p>The position holder will be part of the Corporate Sales US team supporting Front Office sales team, Management, COO through the provision of key client insights and services. You will work closely with the COO office, Corporate Sales regional/country heads and sales leads in onshore locations.</p>\n<p>Your responsibilities will include executing market research and market commentary writ-ups, building actionable intelligence across the corporate client base, maintaining Corporate Sales Marketing content hub, handling global stakeholders, incorporating external market information into the analytics function, preparing pre-meeting client packs for sales team members, and providing commentary on industry trends.</p>\n<p>You will also support Corporate Sales regional heads on ad-hoc as well as Business as Usual (BAU) data and other ad-hoc requests.</p>\n<p>As an HSBC employee, you will have access to tailored professional development opportunities to ensure you have the right skills for today and tomorrow. We offer a competitive pay and benefits package including a robust Wellness Hub, all in a welcoming and inclusive work environment.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_52ea5e8c-da4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610372838","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Direct experience in Banking and Financial Services/Research Companies/Banking Information Technology (IT), Business Analytics, Business Intelligence (BI) Reporting","Well versed on what is currently happening globally about regulations, FX spot, FX forwards, FX options, money market products, interest rates, Swaps and Non-Deliverable Forward (NDF)","Hands on use of Tableau, Alteryx, Qlik Sense or any other visualization tool","Knowledge in VBA, SQL, Python, automation tools","Able to write market commentaries, CCY Pair movement Summaries & Impact of key announcements to currency markets"],"x-skills-preferred":["Expert in Microsoft Office specially in Excel and PowerPoint","Flexibility or adapt to support Asia, Europe, Middle East, and Africa (EMEA), as well as US stakeholders, across different time zones","Able to work independently, proactively and against multiple deadlines"],"datePosted":"2026-04-18T22:09:54.946Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"Direct experience in Banking and Financial Services/Research Companies/Banking Information Technology (IT), Business Analytics, Business Intelligence (BI) Reporting, Well versed on what is currently happening globally about regulations, FX spot, FX forwards, FX options, money market products, interest rates, Swaps and Non-Deliverable Forward (NDF), Hands on use of Tableau, Alteryx, Qlik Sense or any other visualization tool, Knowledge in VBA, SQL, Python, automation tools, Able to write market commentaries, CCY Pair movement Summaries & Impact of key announcements to currency markets, Expert in Microsoft Office specially in Excel and PowerPoint, Flexibility or adapt to support Asia, Europe, Middle East, and Africa (EMEA), as well as US stakeholders, across different time zones, Able to work independently, proactively and against multiple deadlines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f6deb282-e3c"},"title":"Senior Backend Developer (all genders)","description":"<p>Join our Host Experience department as a Senior Backend Developer and become part of the team that powers how our hosts&#39; vacation rentals reach the world.</p>\n<p>You&#39;ll be working at the core of our distribution engine - where we take tens of thousands of homes and make them bookable on major travel platforms such as Holidu, Booking.com, Airbnb, VRBO, HomeToGo, and Check24.</p>\n<p>This team operates in one of the most technically dynamic areas of our product. You will work with systems that synchronize large volumes of updates at high speed and maintain high availability, while integrating with a wide variety of partner APIs - each with its own structure and complexity.</p>\n<p>It&#39;s work that demands precision, scalability, and smart engineering decisions, and it plays a crucial role in helping our hosts reach millions of guests worldwide.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Backend written in Kotlin and Java 21+ (with Spring Boot), with Gradle.</li>\n<li>Deployed as microservices on AWS-hosted Kubernetes cluster (EKS).</li>\n<li>Internal and external web applications written with ReactJS.</li>\n<li>Event-driven communication between services through EventBridge with SQS / ActiveMQ.</li>\n<li>Usage of a diverse set of technologies depending on the use case, such as PostgreSQL, S3, Valkey, ElasticSearch, GraphQL, and many more.</li>\n<li>Monitoring with OpenTelemetry, Grafana, Prometheus, ELK, APM, and CloudWatch.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<ul>\n<li>Design, build, evolve, and maintain our services, creating a great user experience for our hosts.</li>\n<li>Build a strong understanding of the product, use it to drive initiatives end-to-end, and actively shape the team&#39;s direction , not just execute on it.</li>\n<li>Work AI-first: use AI to accelerate not just coding, but data exploration, codebase understanding, technical design, and decision-making , and continuously sharpen how you use these tools.</li>\n<li>Ensure our applications are highly scalable, capable of handling tens of thousands of properties and millions of bookings.</li>\n<li>Work with data persistence - whether in PostgreSQL, Redis, S3, or new state-of-the-art technologies you help us evaluate.</li>\n<li>Ship to production daily , deploying to our AWS Kubernetes cluster is part of the routine, not a special occasion.</li>\n<li>Own the reliability of your services , set up monitoring, define SLOs, and drive incident resolution so your team can move fast with confidence.</li>\n<li>Collaborate in a supportive, cross-functional team that values knowledge sharing and improving together.</li>\n<li>Apply engineering best practices, and stay curious by experimenting with new technologies.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>A passion for great user experience and drive to deliver world-class products.</li>\n<li>Proven track record of delivering product impact through engineering , not just building services, but solving real problems for users.</li>\n<li>Experience with Java or Kotlin with Spring is a plus.</li>\n<li>Experience with relational databases and deploying apps in cloud environments. NoSQL experience is a plus.</li>\n<li>Familiarity with various API types and integration best practices.</li>\n<li>Strong problem-solving skills and a team-oriented mindset.</li>\n<li>Curiosity for the business side - you want to understand the “why” behind the features.</li>\n<li>A love for coding and building high-quality products that make a difference.</li>\n<li>High motivation to learn and experiment with new technologies.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f6deb282-e3c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2573674","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Java","Kotlin","Spring Boot","Gradle","AWS-hosted Kubernetes cluster","ReactJS","EventBridge","SQS","ActiveMQ","PostgreSQL","S3","Valkey","ElasticSearch","GraphQL","OpenTelemetry","Grafana","Prometheus","ELK","APM","CloudWatch"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:09:50.075Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Kotlin, Spring Boot, Gradle, AWS-hosted Kubernetes cluster, ReactJS, EventBridge, SQS, ActiveMQ, PostgreSQL, S3, Valkey, ElasticSearch, GraphQL, OpenTelemetry, Grafana, Prometheus, ELK, APM, CloudWatch"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_39c55814-2f7"},"title":"Manager, Fraud Analytics","description":"<p>In this role, you will drive fraud analytics capability for United States IWPB by owning end-to-end analytic processes. You&#39;ll develop deep expertise in fraud data, detection performance, and client impact, turning insights into measurable reductions in fraud losses and friction while improving controls and decisioning.</p>\n<p>As our Manager, Fraud Analytics, you will:</p>\n<ul>\n<li>Define requirements, manage delivery cadence, document methodologies, and ensure controls/traceability across the analytics lifecycle</li>\n<li>Analyze fraud trends, typologies, and emerging threats using internal/external data to identify root causes and actionable interventions</li>\n<li>Monitor and improve key metrics (e.g., fraud loss rate, detection/true positive rate, false positives, client friction, alert volumes, operational capacity impacts)</li>\n<li>Recommend and test changes to rules, thresholds, segmentation, and model features</li>\n<li>Produce clear recommendations for Fraud Management, Operations, Product, Digital, and Technology, support implementation and post-change validation</li>\n<li>Create concise dashboards and narratives that connect fraud decisions to client experience and business outcomes</li>\n<li>Ensure analyses and changes are well-controlled, auditable, and aligned to relevant policies, model risk expectations, and regulatory considerations</li>\n</ul>\n<p>You&#39;ll likely have the following qualifications to succeed in this role:</p>\n<ul>\n<li>Analytics experience in fraud, financial crime, risk analytics, or payments (banking preferred)</li>\n<li>Advanced capability in SAS/SQL and Python or R</li>\n<li>Proven ability to translate complex analysis into business decisions and influence cross-functional partners without formal authority</li>\n<li>Experience with card, digital payments, digital authentication, account takeover, application fraud, or transaction monitoring</li>\n<li>Dashboarding experience (e.g., QlikSense, Tableau, Power BI)</li>\n<li>Analytical depth, curiosity, and structured problem-solving</li>\n<li>Data storytelling and stakeholder management</li>\n<li>Pragmatic, outcome-driven mindset</li>\n<li>Strong attention to detail and control discipline</li>\n<li>Collaborative working style</li>\n</ul>\n<p>As an HSBC employee, you will have access to tailored professional development opportunities to ensure you have the right skills for today and tomorrow. We offer a competitive pay and benefits package including a robust Wellness Hub, all in a welcoming and inclusive work environment.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_39c55814-2f7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Intl Wealth & Premier Banking","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610398923","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Analytics experience in fraud, financial crime, risk analytics, or payments","Advanced capability in SAS/SQL and Python or R","Proven ability to translate complex analysis into business decisions and influence cross-functional partners without formal authority","Experience with card, digital payments, digital authentication, account takeover, application fraud, or transaction monitoring","Dashboarding experience (e.g., QlikSense, Tableau, Power BI)"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:09:35.474Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"Analytics experience in fraud, financial crime, risk analytics, or payments, Advanced capability in SAS/SQL and Python or R, Proven ability to translate complex analysis into business decisions and influence cross-functional partners without formal authority, Experience with card, digital payments, digital authentication, account takeover, application fraud, or transaction monitoring, Dashboarding experience (e.g., QlikSense, Tableau, Power BI)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7bcb4d82-b90"},"title":"Working Student Backend Engineering (all genders)","description":"<p>You will be working as a Working Student in the Account Compliance &amp; Experience (ACE) team, which is responsible for delivering secure and seamless flows for account lifecycle, relationship, and compliance to customers.</p>\n<p>As a Working Student, you will contribute to the development of new backend features across the ACE domain, assist with operational tasks, get hands-on with modern AI-assisted development, and support ongoing tech refactoring efforts.</p>\n<p>You will work directly alongside senior engineers, take part in real product development, and gradually build ownership over meaningful parts of our codebase.</p>\n<p>The ACE team works within Holidu&#39;s broader backend ecosystem, using Java/Kotlin with Spring Boot, PostgreSQL, Redis, and other data stores, as well as AWS services and Jenkins for CI/CD.</p>\n<p>You will have the opportunity to attend team planning sessions, architecture discussions, and retrospectives, giving you a real window into how a senior engineering team operates in a high-growth company.</p>\n<p>We offer a fair salary, impact, growth, community, flexibility, and fitness opportunities.</p>\n<p>You will be required to work ~20 hours per week, with 1-2 days per week in the office in Munich.</p>\n<p>You should be currently enrolled in a degree in Computer Science, Software Engineering, or a related field, have a solid understanding of object-oriented programming and basic software design principles, and some hands-on experience with Java or Kotlin.</p>\n<p>You should also have familiarity with RESTful APIs and relational databases (SQL), a genuine curiosity for backend systems, and a product-minded attitude.</p>\n<p>Excellent communication skills in English are required, and German is a plus but not required.</p>\n<p>Bonus points if you have exposure to Spring Boot, cloud platforms (AWS), or any experience with identity/access management concepts.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7bcb4d82-b90","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2605407","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"working_student","x-salary-range":null,"x-skills-required":["Java","Kotlin","Spring Boot","PostgreSQL","Redis","AWS services","Jenkins","CI/CD","RESTful APIs","relational databases (SQL)"],"x-skills-preferred":["cloud platforms (AWS)","identity/access management concepts"],"datePosted":"2026-04-18T22:09:31.209Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Kotlin, Spring Boot, PostgreSQL, Redis, AWS services, Jenkins, CI/CD, RESTful APIs, relational databases (SQL), cloud platforms (AWS), identity/access management concepts"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_80d15de9-aa7"},"title":"Senior Data Scientist - Rankings & Recommendations (all genders)","description":"<p>Join our Business Intelligence Department, a multidisciplinary group of Data Scientists, Analysts, and Data Engineers.</p>\n<p>You will join a cross-functional Product team, Search Intelligence, which is responsible for optimizing ranking and recommendations for users visiting our website.</p>\n<p>You&#39;ll be part of the broader Data Science team, which operates across cross-functional domain teams - giving you access to shared knowledge, best practices, and collaboration opportunities beyond your domain.</p>\n<p>You’ll collaborate daily with Data Engineers, Analysts, Product Managers, and Back-end Engineers.</p>\n<p>You’ll report to the Team Lead, Data Science.</p>\n<p>Together, we turn data into actionable insights and innovative technology that powers how millions of guests find and book their perfect holiday home.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Python • Airflow • dbt • AWS (SageMaker, Redshift, Athena) • MLflow</li>\n</ul>\n<p>The Ranking challenge at Holidu</p>\n<p>Holidu lists over 4 million vacation rental properties. Our ranking and personalization systems determine which of them our 70+ million annual users see, directly impacting search conversion and business results.</p>\n<p>What&#39;s live today:</p>\n<ul>\n<li>Multi-stage ranking pipeline: Reinforcement-learning-based cold ranking, contextual re-ranking, and personalized recommendations.</li>\n</ul>\n<ul>\n<li>Cold-start models for new properties with limited behavioral data.</li>\n</ul>\n<ul>\n<li>Personalized recommendations based on user browsing patterns.</li>\n</ul>\n<p>Some of the hard problems we&#39;re solving:</p>\n<ul>\n<li>Multi-objective optimization: Balancing user relevance, conversion probability, and business value.</li>\n</ul>\n<ul>\n<li>Personalization without history: Most users are anonymous or first-time visitors.</li>\n</ul>\n<ul>\n<li>Cold-start: A significant share of our inventory is new each quarter. How do we surface quality properties before we have behavioral data?</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>You&#39;ll shape the ranking and recommendation systems that millions of guests rely on to find their holiday home. With access to extensive datasets and modern ML infrastructure, you&#39;ll work end-to-end - from identifying opportunities and prototyping new approaches to shipping models to production and measuring their impact.</p>\n<ul>\n<li>Develop high-impact models and improvements for our ranking, recommendation, and personalization systems - with the freedom to explore new, creative approaches.</li>\n</ul>\n<ul>\n<li>Take models from conception to production, continuously monitor their performance, and iterate to enhance accuracy and efficiency.</li>\n</ul>\n<ul>\n<li>Design and run A/B tests as a core part of ranking development; success is measured by successful experiments per quarter and time-to-decision.</li>\n</ul>\n<ul>\n<li>Collaborate closely with Product Managers and Software Engineers to identify, prioritize, and ship ranking improvements.</li>\n</ul>\n<ul>\n<li>Ensure model reliability in production, measured by online/offline agreement, model and data drift KPIs, latency and uptime SLAs, and automated monitoring coverage.</li>\n</ul>\n<ul>\n<li>Advance our MLOps practices with CI/CD pipelines, retraining workflows, lineage tracking, and documentation.</li>\n</ul>\n<ul>\n<li>Demonstrate leadership in data science projects by driving technical direction, scoping initiatives, and guiding the team&#39;s prioritization and project execution.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>5+ years of experience as a Data Scientist, with a proven track record of applying ML models to solve real business problems.</li>\n</ul>\n<ul>\n<li>Experience working on ranking models or recommender systems is a strong advantage.</li>\n</ul>\n<ul>\n<li>A degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field.</li>\n</ul>\n<ul>\n<li>Strong foundations in statistics, predictive modeling, and machine learning techniques, with hands-on experience using Python and SQL.</li>\n</ul>\n<ul>\n<li>Experience with Airflow and dbt is a plus.</li>\n</ul>\n<ul>\n<li>Solid understanding of business operations and the ability to translate data insights into clear, actionable outcomes.</li>\n</ul>\n<ul>\n<li>A collaborative mindset and enthusiasm for using data to build world-class products that make a real impact.</li>\n</ul>\n<ul>\n<li>AI Proficiency: You are comfortable using AI to enhance coding, planning, and monitoring. This includes successfully integrating AI tools (such as Claude code, Codex, Copilot, etc.) into your workflow and teaching others to use them efficiently.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n</ul>\n<ul>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n</ul>\n<ul>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n</ul>\n<ul>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n</ul>\n<ul>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n</ul>\n<ul>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p>Need a sneak peek? Check out the adventure that awaits you on Instagram @lifeatholidu and dive straight into the world of Tech at Holidu for more insights!</p>\n<p><strong>Want to travel with us?</strong></p>\n<p>Apply online on our careers page! Your first travel contact will be Lucia from HR.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_80d15de9-aa7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2413808","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","Airflow","dbt","AWS","MLflow","Machine Learning","Statistics","Predictive Modeling","SQL"],"x-skills-preferred":["AI","Data Science","Ranking Models","Recommender Systems","Collaboration","Communication"],"datePosted":"2026-04-18T22:09:15.403Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Airflow, dbt, AWS, MLflow, Machine Learning, Statistics, Predictive Modeling, SQL, AI, Data Science, Ranking Models, Recommender Systems, Collaboration, Communication"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f3494283-f86"},"title":"Strategic Finance Manager, International Public Sector","description":"<p>We are seeking a Strategic Finance Manager to join our CFO Team in London, UK. As a key member of our Finance team, you will play a crucial role in supporting the G&amp;A and EPD leadership in quarterly and annual strategic and financial planning discussions. You will engage with the business to manage headcount and expense forecasts, own weekly and monthly financial reporting, KPIs, presentations, and other critical analysis requests. You will partner with Business Unit Owners to own and improve Hosting Infrastructure expense forecasting. Your responsibilities will also include evaluating investment strategies, facilitating ad-hoc analyses, and owning special projects requiring complex financial models. You will drive internal initiatives to improve data infrastructure and reporting, and partner closely with the accounting team during the month-end close process to ensure accurate general ledger and company-wide financial statement reporting. You will develop and maintain KPIs to assist with the decision-making that aligns with the company&#39;s objectives.</p>\n<p>Ideally, you will have 6+ years of financial analysis experience working directly in a Corporate Finance or a Strategic Finance role, with 2 years of investment banking or top-tier consulting experience. You will have demonstrated excellent project management skills, ability to manage and manipulate large data sets, critically analyze existing processes, and identify opportunities for process improvement. You will have strong knowledge of Google Suite, and MS Office (Excel, Word, PowerPoint); expert Excel modeling skills.</p>\n<p>The ideal candidate will be a strong communicator with excellent interpersonal skills to manage various stakeholders. You will be able to work closely with department heads on real-time high-priority business issues and use quantitative insights to drive better decision-making across the organization.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f3494283-f86","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4631613005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["financial analysis","strategic finance","investment banking","top-tier consulting","project management","Google Suite","MS Office","Excel modeling"],"x-skills-preferred":["SQL","Business Intelligence tools","Anaplan","Adaptive Insights"],"datePosted":"2026-04-18T16:02:14.471Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Technology","skills":"financial analysis, strategic finance, investment banking, top-tier consulting, project management, Google Suite, MS Office, Excel modeling, SQL, Business Intelligence tools, Anaplan, Adaptive Insights"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_413284d2-da2"},"title":"Strategic Finance Manager, Gen AI","description":"<p>We&#39;re building out the Finance team to help make data-driven and financially sound decisions for Scale. The Finance team drives strategic, financial, and operational decisions by partnering with the leadership team to make critical decisions across Scale.</p>\n<p>We’re looking for a high-performing, all-rounded finance athlete to join our team and support the rapidly growing Generative AI (GenAI) business. You’ll collaborate closely with Product, Operations, Growth, and Go-to-Market leaders to bring financial rigor to decision-making, develop actionable insights that drive strategy, and build scalable systems as the business expands.</p>\n<p>This role is ideal for someone with 4-6 years of experience in a fast-paced, high-growth environment. Someone who thrives in ambiguity, can juggle multiple workstreams, and brings a mix of analytical rigor, business acumen, and strong execution.</p>\n<p>You will:</p>\n<ul>\n<li>Own and evolve part of the GenAI financial forecasting model, driving accuracy and insight across planning cycles</li>\n<li>Support reporting and performance management, including weekly and monthly reviews, consolidations, and ad hoc analyses</li>\n<li>Partner with GenAI leadership and cross functional teams to evaluate and execute key strategic and operational initiatives that scale the business multifold</li>\n<li>Conduct financial analyses and build business cases for new products, partnerships, and investments</li>\n<li>Collaborate with Accounting, and Corporate Finance to improve close, reporting, and planning cadences</li>\n<li>Continuously improve financial processes and systems to enhance scalability, forecast precision, and data visibility</li>\n</ul>\n<p>Ideally, you&#39;d have:</p>\n<ul>\n<li>4–6 years of experience in Strategic Finance, FP&amp;A, or Business Operations, ideally within a high-growth technology company</li>\n<li>2 years of investment banking experience at a top-tier firm</li>\n<li>Strong analytical and financial modeling skills; ability to translate complex data into actionable insights</li>\n<li>Excellent communication skills, with the ability to distill complexity into clear narratives for non-finance stakeholders</li>\n<li>Advanced proficiency in Excel, Google Sheets, and PowerPoint; strong command of financial modeling best practices</li>\n<li>Experience with SQL or Business Intelligence tools (e.g., Looker, Tableau)</li>\n<li>Familiarity with Anaplan, Adaptive Insights, or other planning systems</li>\n</ul>\n<p>Nice to haves:</p>\n<ul>\n<li>Bachelor’s degree in Finance, Accounting, Economics, Engineering, or a related field</li>\n<li>Prior experience supporting Product, Engineering, Growth, or Operations teams within a technology company</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_413284d2-da2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4623533005","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$176,400-$220,500 USD","x-skills-required":["financial modeling","strategic finance","FP&A","business operations","investment banking","Excel","Google Sheets","PowerPoint","SQL","Business Intelligence","Anaplan","Adaptive Insights"],"x-skills-preferred":["data analysis","business acumen","communication","leadership"],"datePosted":"2026-04-18T16:01:35.646Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Technology","skills":"financial modeling, strategic finance, FP&A, business operations, investment banking, Excel, Google Sheets, PowerPoint, SQL, Business Intelligence, Anaplan, Adaptive Insights, data analysis, business acumen, communication, leadership","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":176400,"maxValue":220500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_90b5ac1d-d16"},"title":"Senior Software Engineer, Backend — Frontier Data","description":"<p>The Frontier Data team builds the data and systems that power Scale&#39;s most advanced Frontier AI use cases. We&#39;re looking for a Senior Backend Engineer who thrives in ambiguity, moves fast, and enjoys tackling daunting challenges.</p>\n<p>As a Senior Backend Engineer, you will own major backend systems for frontier agentic data products, driving projects from early exploration through production deployment. You will build scalable services and pipelines that support agent workflows, architect modular, reusable backend systems, and operate in high-ambiguity environments.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Designing and building scalable systems while partnering closely with research, product, operations, and other engineering teams</li>\n<li>Building scalable services and pipelines that support agent workflows</li>\n<li>Architecting modular, reusable backend systems that adapt to evolving product needs</li>\n<li>Operating in high-ambiguity environments and breaking down open-ended problems</li>\n<li>Partnering cross-functionally with product, research/ML, and infrastructure teams</li>\n</ul>\n<p>Ideal experience includes 5+ years of full-time software engineering experience, strong backend engineering fundamentals, and experience building systems that scale.</p>\n<p>Compensation packages at Scale include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors.</p>\n<p>Additional benefits include comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_90b5ac1d-d16","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Frontier Data","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4648525005","x-work-arrangement":null,"x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["Distributed systems","API design","Data modeling","Production reliability","Docker","Containerized development/production environments","SQL","Modern database-backed application development"],"x-skills-preferred":["Async processing","Workflow engines","Data pipelines"],"datePosted":"2026-04-18T16:01:34.567Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Distributed systems, API design, Data modeling, Production reliability, Docker, Containerized development/production environments, SQL, Modern database-backed application development, Async processing, Workflow engines, Data pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_38e51e8f-9b2"},"title":"Lead Technical Program Manager, Trust & Safety","description":"<p>About the Role</p>\n<p>Scale is at the frontier of GenAI and human-AI collaboration. The Gen AI Ops Trust and Safety team is focused on safeguarding human authenticity and genuineness in AI training.</p>\n<p>We are looking for a highly analytical Technical Program Manager (TPM) who leans heavily into fraud analytics and data-driven strategy to protect our ecosystem. This isn&#39;t a project management role. You will act as the lead investigative analyst and program owner for our fraud defense portfolio.</p>\n<p>Your day-to-day will involve diving deep into complex datasets to uncover hidden fraud vectors, and then translating those analytical insights into scalable rules, policies, and operational programs. By utilizing AI coding tools at high velocity, you will build out analytics pipelines, dashboards, and detection logic to shift Trust and Safety from a reactive function to a strategic one that balances safety and growth.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Analyze large, messy behavioural events to identify ambiguous and constantly evolving fraud patterns across the contributor lifecycle.</li>\n</ul>\n<ul>\n<li>Translate your analytical findings into actionable detection logic. You will redesign rules, optimize thresholds, and decision flows to catch bad actors while minimizing friction for high-quality contributors.</li>\n</ul>\n<ul>\n<li>Establish robust KPIs, build tracking dashboards, and define offline evaluation frameworks (e.g., false positive monitoring, precision/recall analysis) to continuously measure the health of our risk strategy.</li>\n</ul>\n<ul>\n<li>Act as the connective tissue between data, operations, and engineering. You will use your analytical findings to implement technical execution, taking new detection capabilities from data prototype to production deployment.</li>\n</ul>\n<ul>\n<li>Leverage AI-assisted IDEs daily to rapidly write complex SQL queries, automate data pulls, and streamline your analytical workflows.</li>\n</ul>\n<ul>\n<li>Connect signals, data, and operations to see the full picture. Provide clear, direct communication regarding fraud trends and strategy shifts to both technical and non-technical partners.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>5-8 years of experience in risk strategy and fraud analytics. You have a battle-tested track record of reverse-engineering adversarial patterns, dismantling complex fraud vectors, and driving highly analytical Trust &amp; Safety or TPM programs.</li>\n</ul>\n<ul>\n<li>Expert proficiency in SQL skills. You must be highly comfortable extracting insights from large datasets in noisy, adversarial environments.</li>\n</ul>\n<ul>\n<li>An execution-driven mindset focused on delivering measurable results, not just theoretical analysis. You are comfortable working in ambiguity and taking ownership from data problems to operational solutions.</li>\n</ul>\n<ul>\n<li>Strong proficiency with AI coding assistants to accelerate data exploration and query writing.</li>\n</ul>\n<ul>\n<li>Deep understanding of how to balance aggressive fraud detection with marketplace growth. You make decisions based on what&#39;s right for the business, not what&#39;s convenient.</li>\n</ul>\n<p>Nice to haves:</p>\n<ul>\n<li>Solid Python skills (e.g., Pandas, NumPy) for advanced data manipulation and scripting</li>\n</ul>\n<ul>\n<li>Experience working in marketplace or gig-economy platforms is a plus</li>\n</ul>\n<p>Compensation</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>Salary Range</p>\n<p>The base salary range for this full-time position in the location of San Francisco is: $180,800-$226,000 USD</p>\n<p>Benefits</p>\n<p>You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p>About Us</p>\n<p>At Scale, our mission is to develop reliable AI systems for the world&#39;s most important decisions. Our products provide the high-quality data and full-stack technologies that power the world&#39;s leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact.</p>\n<p>We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.</p>\n<p>We believe that everyone should be able to bring their whole selves to work, which is why we are proud to be an inclusive and equal opportunity workplace. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability status, gender identity or Veteran status.</p>\n<p>We are committed to working with and providing reasonable accommodations to applicants with physical and mental disabilities. If you need assistance and/or a reasonable accommodation in the application or recruiting process due to a disability, please contact us at accommodations@scale.com.</p>\n<p>Please see the United States Department of Labor&#39;s Know Your Rights poster for additional information.</p>\n<p>We comply with the United States Department of Labor&#39;s Pay Transparency provision.</p>\n<p>PLEASE NOTE: We collect, retain and use personal data for our professional business purposes, including notifying you of job opportunities that may be of interest and sharing with our affiliates. We limit the personal data we collect to that which we believe is appropriate and necessary to manage applicants’ needs, provide our services, and comply with applicable laws. Any information we collect in connection with your application will be treated in accordance with our internal policies and programs designed to protect personal data. Please see our privacy policy for additional information.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_38e51e8f-9b2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4674924005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,800-$226,000 USD","x-skills-required":["fraud analytics","data-driven strategy","SQL skills","AI coding assistants","Python skills"],"x-skills-preferred":["Pandas","NumPy","experience working in marketplace or gig-economy platforms"],"datePosted":"2026-04-18T16:01:32.183Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"fraud analytics, data-driven strategy, SQL skills, AI coding assistants, Python skills, Pandas, NumPy, experience working in marketplace or gig-economy platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180800,"maxValue":226000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c19e39af-feb"},"title":"Full-Stack Software Engineer, (Forward Deployed), GPS","description":"<p>Scale&#39;s rapidly growing Global Public Sector team is focused on using AI to address critical challenges facing the public sector around the world.</p>\n<p>Our core work consists of creating custom AI applications that will impact millions of citizens, generating high-quality training data for custom LLMs, and upskilling and advisory services to spread the impact of AI.</p>\n<p>As a Full Stack Software Engineer (Forward Deployed), you&#39;ll collaborate directly with public sector counterparts to quickly build full-stack, AI applications, to solve their most pressing challenges and achieve meaningful impact for citizens.</p>\n<p>At Scale, we&#39;re not just building AI solutions,we&#39;re enabling the public sector to transform their operations and better serve citizens through cutting-edge technology.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Collaborate with senior engineers to implement features for public sector clients, including spending time with the client to understand user feedback and assist with delivery.</li>\n<li>Develop and maintain full-stack components that integrate with AI models, focusing on building responsive UIs and reliable backend APIs.</li>\n<li>Assist in deploying and monitoring applications within cloud environments, ensuring basic system stability and security.</li>\n<li>Help build and refine reusable features that support diverse international client use cases.</li>\n<li>Work within a multi-disciplinary team of design, product, and data specialists to build robust features that follow established technical architectures.</li>\n</ul>\n<p><strong>Ideal Candidate:</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science or a related quantitative field</li>\n<li>Professional full-stack experience with a focus on React, TypeScript, and Python/Node.js. Familiarity with Next.js and NoSQL/Relational databases, along with exposure to containerization (Docker) and cloud deployments.</li>\n<li>Experience building and deploying web applications with a good understanding of cloud fundamentals and scalable coding practices.</li>\n<li>A self-starting approach to navigate ambiguous requirements and deliver reliable software.</li>\n</ul>\n<p><strong>Nice to Have:</strong></p>\n<ul>\n<li>Proficient in Arabic</li>\n<li>Experience working cross functionally with operations</li>\n<li>Experience building solutions with LLMs</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c19e39af-feb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4676602005","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["React","TypeScript","Python","Node.js","Next.js","NoSQL/Relational databases","containerization (Docker)","cloud deployments"],"x-skills-preferred":["Arabic","experience working cross functionally with operations","experience building solutions with LLMs"],"datePosted":"2026-04-18T16:01:21.167Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dubai, UAE"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Python, Node.js, Next.js, NoSQL/Relational databases, containerization (Docker), cloud deployments, Arabic, experience working cross functionally with operations, experience building solutions with LLMs"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2d16873c-e17"},"title":"Full-Stack Software Engineer, (Forward Deployed), GPS","description":"<p>Scale&#39;s rapidly growing Global Public Sector team is focused on using AI to address critical challenges facing the public sector around the world.</p>\n<p>Our core work consists of creating custom AI applications that will impact millions of citizens, generating high-quality training data for custom LLMs, and upskilling and advisory services to spread the impact of AI.</p>\n<p>As a Full Stack Software Engineer (Forward Deployed), you&#39;ll collaborate directly with public sector counterparts to quickly build full-stack, AI applications, to solve their most pressing challenges and achieve meaningful impact for citizens.</p>\n<p>At Scale, we&#39;re not just building AI solutions,we&#39;re enabling the public sector to transform their operations and better serve citizens through cutting-edge technology.</p>\n<p>If you&#39;re ready to shape the future of AI in the public sector and be a founding member of our team, we&#39;d love to hear from you.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Collaborate with senior engineers to implement features for public sector clients, including spending time with the client to understand user feedback and assist with delivery.</li>\n<li>Develop and maintain full-stack components that integrate with AI models, focusing on building responsive UIs and reliable backend APIs.</li>\n<li>Assist in deploying and monitoring applications within cloud environments, ensuring basic system stability and security.</li>\n<li>Help build and refine reusable features that support diverse international client use cases.</li>\n<li>Work within a multi-disciplinary team of design, product, and data specialists to build robust features that follow established technical architectures.</li>\n</ul>\n<p><strong>Ideal Candidate</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science or a related quantitative field</li>\n<li>Professional full-stack experience with a focus on React, TypeScript, and Python/Node.js. Familiarity with Next.js and NoSQL/Relational databases, along with exposure to containerization (Docker) and cloud deployments.</li>\n<li>Experience building and deploying web applications with a good understanding of cloud fundamentals and scalable coding practices.</li>\n<li>A self-starting approach to navigate ambiguous requirements and deliver reliable software.</li>\n</ul>\n<p><strong>Nice to Haves</strong></p>\n<ul>\n<li>Proficient in Arabic</li>\n<li>Experience working cross functionally with operations</li>\n<li>Experience building solutions with LLMs</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2d16873c-e17","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4676600005","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["React","TypeScript","Python","Node.js","Next.js","NoSQL/Relational databases","containerization (Docker)","cloud deployments"],"x-skills-preferred":["Arabic","cross functional collaboration","LLM solutions"],"datePosted":"2026-04-18T16:01:13.044Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Doha, Qatar"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Python, Node.js, Next.js, NoSQL/Relational databases, containerization (Docker), cloud deployments, Arabic, cross functional collaboration, LLM solutions"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d30384aa-64a"},"title":"Strategic Projects Lead - Coding","description":"<p>Scale&#39;s Generative AI business unit is experiencing historic levels of growth. As a Strategic Projects Lead, you will lead initiatives that drive $XXM+ in new revenue for the business. This is a demanding role, requiring a strong entrepreneurial mindset, comfort with getting into the weeds, and excitement about intense, impactful work that leads to accelerated career progression.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Serve as the full owner of our most visible and high-impact customer pipelines, making decisions that directly impact data quality, operational efficiency, revenue, and margins</li>\n<li>Understand customer requirements and design data taxonomy best suited to improving model performance based on customer needs</li>\n<li>Build out pipeline infrastructure to ensure quality and efficiency</li>\n<li>Train, coach, and manage dynamic and global teams</li>\n<li>Build analytics to make data-driven decisions</li>\n<li>Partner with diverse stakeholders (Engineering + Product + Ops + Go-to-Market) to work on problems that will drive advancements for the largest LLMs in the world</li>\n<li>Give regular progress updates to Scale&#39;s executive team</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>A strong technical background is required for this position. For example, a degree in Machine Learning Engineering, Computer Science, or Software Engineering</li>\n<li>3+ years of experience leading a team/projects, managing operational processes, or 3+ years of experience as a SWE</li>\n<li>Strong problem-solving capabilities in technical environments</li>\n<li>Ability to come up with creative solutions to complex, ambiguous, operational, and technical problems</li>\n<li>Entrepreneurial experience and mindset</li>\n</ul>\n<p>Compensation and Benefits: The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity-based compensation, subject to Board of Director approval. You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d30384aa-64a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4666036005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$112,000-$190,000 USD","x-skills-required":["Machine Learning Engineering","Computer Science","Software Engineering","SQL","Python","Data Analytics"],"x-skills-preferred":[],"datePosted":"2026-04-18T16:00:51.409Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning Engineering, Computer Science, Software Engineering, SQL, Python, Data Analytics","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":112000,"maxValue":190000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3cc878fa-5d1"},"title":"Infrastructure Software Engineer, Enterprise GenAI","description":"<p>We are seeking a strong engineer to join our team and help us build and scale our core infrastructure in a fast-paced environment. The ideal candidate will have a strong understanding of software engineering principles and practices, as well as experience with large-scale distributed systems.</p>\n<p>You will implement solutions across multiple cloud providers (GCP, Azure, AWS) for customers in diverse, highly-regulated industries like healthcare, telecom, finance, and retail.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Architecting multi-cloud systems and abstractions to allow the SGP platform to run on top of existing Cloud providers</li>\n<li>Implementing custom integrations between Scale AI&#39;s platform and customer data environments (cloud platforms, data warehouses, internal APIs)</li>\n<li>Collaborating with platform, product teams and our customers directly to develop and implement innovative infrastructure that scales to meet evolving needs</li>\n<li>Delivering experiments at a high velocity and level of quality to engage our customers</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>4+ years of full-time engineering experience, post-graduation</li>\n<li>Experience scaling products at hyper growth startups</li>\n<li>Experience tinkering with or productizing LLMs, vector databases, and the other latest AI technologies</li>\n<li>Proficient in Python or Javascript/Typescript, and SQL</li>\n<li>Experience with Kubernetes</li>\n<li>Experience with major cloud providers (AWS, Azure, GCP)</li>\n<li>Excellent communication skills with the ability to explain technical concepts to both technical and non-technical audiences</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3cc878fa-5d1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4665557005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$179,400-$224,250 USD","x-skills-required":["Python","Javascript/Typescript","SQL","Kubernetes","GCP","Azure","AWS"],"x-skills-preferred":[],"datePosted":"2026-04-18T16:00:45.380Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Javascript/Typescript, SQL, Kubernetes, GCP, Azure, AWS","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":179400,"maxValue":224250,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_012ee214-ce3"},"title":"Quality Lead II","description":"<p>We&#39;re looking for a Quality Lead to ensure top-tier data quality across our Generative AI programs. As a Quality Lead, you will be the authority on the usefulness of the data we send to our customers. You will hold delivery teams accountable across our highest-impact projects, engaging in nuanced quality issues head-on, as well as leading teams through strategic initiatives to enable reliable, scalable delivery across the organisation.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Leading and developing a team of quality control specialists</li>\n<li>Partnering with delivery teams to interpret and improve customer requirements</li>\n<li>Overseeing day-to-day operations, including workload allocation and reporting insights</li>\n<li>Establishing and upholding scalable best practices</li>\n<li>Driving initiatives that improve quality, efficiency, and customer satisfaction</li>\n</ul>\n<p>Ideally, you&#39;d have:</p>\n<ul>\n<li>A Bachelor&#39;s degree in Computer Science, Engineering, Operations, or a related field</li>\n<li>4+ years leading teams or improving operational/technical processes</li>\n<li>Strong analytical and problem-solving skills (particularly in SQL) with attention to detail</li>\n<li>Experience building systems, processes, and teams across multiple customer segments or product lines</li>\n<li>A hands-on background in quality assurance, software development, or top-tier management consulting</li>\n<li>Proven experience taking initiative in cross-functional technical environments</li>\n</ul>\n<p>Nice to haves:</p>\n<ul>\n<li>Master&#39;s degree in Computer Science, Engineering, Operations, or a related field</li>\n<li>Experience in AI/ML services</li>\n<li>Experience in a fast-growing, less-structured startup environment</li>\n<li>Experience building teams from the ground up</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_012ee214-ce3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4683041005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$134,400-$168,000 USD","x-skills-required":["SQL","quality assurance","software development","management consulting","team leadership"],"x-skills-preferred":["AI/ML services","fast-growing startup environment","team building"],"datePosted":"2026-04-18T16:00:43.029Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, quality assurance, software development, management consulting, team leadership, AI/ML services, fast-growing startup environment, team building","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":134400,"maxValue":168000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_01e7c642-a6c"},"title":"Engagement Manager, Public Sector","description":"<p>We&#39;re hiring an engagement manager to lead and coordinate delivery of agentic workflows for a national security customer. This role is ideal for someone who blends program leadership, customer relationship building, technical fluency, and contract awareness.</p>\n<p>Key responsibilities include: Managing customer relationships from the executive to the end user Working alongside customers to scope agentic workflow use cases that Scale&#39;s engineering team will build and refine Leading a cross-functional project team to deliver on and exceed the customer&#39;s AI/ML objectives Overseeing onboarding and successful implementation of customer accounts</p>\n<p>Must haves: Active TS/SCI clearance 5+ years of work experience succeeding in stakeholder management or a customer-facing role delivering enterprise-scale applications/solutions A track record of structured, analytics-driven problem solving Excellent verbal and written communication skills Willingness to be onsite with the customer in the Colorado Springs area 4 days per week and able to travel at least 25% of the time</p>\n<p>Compensation packages at Scale include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_01e7c642-a6c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4667833005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$195,800-$279,400 USD","x-skills-required":["customer relationship building","technical fluency","contract awareness","structured problem solving","excellent communication skills"],"x-skills-preferred":["Python","SQL","API technology","domain expertise"],"datePosted":"2026-04-18T16:00:39.776Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Colorado Springs, CO"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"customer relationship building, technical fluency, contract awareness, structured problem solving, excellent communication skills, Python, SQL, API technology, domain expertise","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":195800,"maxValue":279400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_94999453-111"},"title":"Senior Full-Stack Software Engineer, (Forward Deployed), GPS","description":"<p>Scale&#39;s rapidly growing Global Public Sector team is focused on using AI to address critical challenges facing the public sector around the world.</p>\n<p>Our core work consists of creating custom AI applications that will impact millions of citizens, generating high-quality training data for custom LLMs, and upskilling and advisory services to spread the impact of AI.</p>\n<p>As a Full Stack Software Engineer (Forward Deployed), you&#39;ll collaborate directly with public sector counterparts to quickly build full-stack, AI applications, to solve their most pressing challenges and achieve meaningful impact for citizens.</p>\n<p>At Scale, we&#39;re not just building AI solutions,we&#39;re enabling the public sector to transform their operations and better serve citizens through cutting-edge technology.</p>\n<p>If you&#39;re ready to shape the future of AI in the public sector and be a founding member of our team, we&#39;d love to hear from you.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Partner with public sector clients to scope, collect feedback and implement solutions for complex problems, including spending up to two weeks per month in client offices for feedback and delivery.</li>\n<li>Architect production-grade applications that integrate AI models with full-stack frameworks, managing everything from interactive UIs to backend APIs and systems.</li>\n<li>Deploy and manage infrastructure within cloud environments, ensuring the highest levels of system integrity, security, scalability, and long-term reliability.</li>\n<li>Contribute to core platform features designed to be reused across diverse international client use cases.</li>\n<li>Partner with design, product, and data teams to build robust applications aligned with the broader technical architecture.</li>\n</ul>\n<p><strong>Ideal Candidate</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science or a related quantitative field</li>\n<li>5+ years of post-graduation, full-stack engineering experience with demonstrated proficiency in React (required), TypeScript, Next.js, Python, Node.js, PostgreSQL or MongoDB plus hands-on experience with Docker, Kubernetes, and Azure/AWS/GCP.</li>\n<li>Proven ability to architect scalable, production-grade applications with a strong handle on cloud environments and infrastructure health.</li>\n<li>Experience working directly within customer infrastructure to deploy, maintain, and troubleshoot complex, end-to-end solutions.</li>\n<li>A self-starting approach with the technical maturity to navigate ambiguous requirements and deliver reliable software.</li>\n<li>Driven async communication methodologies to reduce communication frictions</li>\n</ul>\n<p><strong>Nice to Haves</strong></p>\n<ul>\n<li>Proficient in Arabic</li>\n<li>Past experience working in a forward deployed engineer / dedicated customer engineer role</li>\n<li>Experience working cross functionally with operations</li>\n<li>Experience building solutions with LLMs and a deep understanding of the overall Gen AI landscape</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_94999453-111","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4676608005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["React","TypeScript","Next.js","Python","Node.js","PostgreSQL","MongoDB","Docker","Kubernetes","Azure","AWS","GCP"],"x-skills-preferred":[],"datePosted":"2026-04-18T16:00:24.081Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dubai, UAE"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Next.js, Python, Node.js, PostgreSQL, MongoDB, Docker, Kubernetes, Azure, AWS, GCP"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_fa90b2bf-155"},"title":"Quality Lead","description":"<p>We&#39;re looking for a Quality Lead to ensure top-tier data quality across our Generative AI programs. In this role, you will partner directly with Scale&#39;s Delivery teams to ensure that the Quality Team&#39;s core operation runs smoothly and that only high-quality data is delivered to our customers.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Leading and developing teams of quality control auditors</li>\n<li>Partnering with delivery teams to interpret and improve customer requirements</li>\n<li>Overseeing day-to-day operations, including workload allocation and reporting insights</li>\n<li>Establishing and upholding scalable best practices</li>\n<li>Driving initiatives that improve quality, efficiency, and customer satisfaction</li>\n</ul>\n<p>Ideal candidates will have a Bachelor&#39;s degree in Computer Science, Engineering, Operations, or a related field, and 2+ years improving operational/technical processes. Strong analytical and problem-solving skills, particularly in SQL and Google Sheets/Excel, are also required.</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>Benefits include comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additional benefits may be available, including a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_fa90b2bf-155","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4667108005","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$109,600-$137,000 USD","x-skills-required":["SQL","Google Sheets/Excel","quality control","process improvement","team leadership"],"x-skills-preferred":["product management","software development","management consulting","AI/ML services","quality assurance"],"datePosted":"2026-04-18T16:00:14.449Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Google Sheets/Excel, quality control, process improvement, team leadership, product management, software development, management consulting, AI/ML services, quality assurance","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":109600,"maxValue":137000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1bebb6dc-380"},"title":"Staff Software Engineer, Platform","description":"<p>We live in unprecedented times – AI has the potential to exponentially augment human intelligence. As the world adjusts to this new reality, leading platform companies are scrambling to build LLMs at billion scale, while large enterprises figure out how to add it to their products.</p>\n<p>At Scale, our products include the Generative AI Data Engine, SGP, Donovan, and others that power the most advanced LLMs and generative models in the world through world-class RLHF, human data generation, model evaluation, safety, and alignment.</p>\n<p>As a Staff Software Engineer, you will define and drive both the architectural roadmap and implementation of core platforms and software systems. You will be responsible for providing high-level vision and driving adoption across the engineering org for orchestration, data abstraction, data pipelines, identity &amp; access management, and underlying cloud infrastructure.</p>\n<p>Impact and Responsibilities:</p>\n<ul>\n<li>Architectural Vision: You will drive the design and implementation of foundational systems, acting as a bridge between high-level business goals and technical goals.</li>\n</ul>\n<ul>\n<li>Cross-Functional Leadership: You will collaborate with cross-functional teams to define and drive adoption of the next generation of features for our AI data infrastructure.</li>\n</ul>\n<ul>\n<li>Technical Ownership: You are responsible for proactively identifying and driving opportunities for organizational growth, driving improvements in programming practices, and upgrading the tools that define our development lifecycle.</li>\n</ul>\n<ul>\n<li>Technical Mentorship: You will serve as a subject matter expert, presenting technical information to stakeholders and providing the guidance to elevate the engineering culture across the company.</li>\n</ul>\n<p>Ideally you’d have:</p>\n<ul>\n<li>8+ years of full-time engineering experience, post-graduation with specialities in back-end systems.</li>\n</ul>\n<ul>\n<li>Extensive experience in software development and a deep understanding of distributed systems and public cloud platforms (AWS preferred).</li>\n</ul>\n<ul>\n<li>Demonstrated a track record of independent ownership and leadership across successful multi-team engineering projects.</li>\n</ul>\n<ul>\n<li>Possess excellent communication and collaboration skills, and the ability to translate complex technical concepts to non-technical stakeholders.</li>\n</ul>\n<ul>\n<li>Experience working fluently with standard containerization &amp; deployment technologies like Kubernetes, Terraform, Docker, etc.</li>\n</ul>\n<ul>\n<li>Experience with orchestration platforms, such as Temporal and AWS Step Functions.</li>\n</ul>\n<ul>\n<li>Experience with NoSQL document databases (MongoDB) and structured databases (Postgres).</li>\n</ul>\n<ul>\n<li>Strong knowledge of software engineering best practices and CI/CD tooling (CircleCI, ArgoCD).</li>\n</ul>\n<p>Nice to haves:</p>\n<ul>\n<li>Experience with data warehouses (Snowflake, Firebolt) and data pipeline/ETL tools (Dagster, dbt).</li>\n</ul>\n<ul>\n<li>Experience scaling products at hyper-growth startups.</li>\n</ul>\n<ul>\n<li>Excitement to work with AI technologies.</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is: $252,000-$315,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1bebb6dc-380","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4649893005","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$252,000-$315,000 USD","x-skills-required":["Software development","Distributed systems","Public cloud platforms","Containerization & deployment technologies","Orchestration platforms","NoSQL document databases","Structured databases","Software engineering best practices","CI/CD tooling"],"x-skills-preferred":["Data warehouses","Data pipeline/ETL tools","Scaling products at hyper-growth startups","AI technologies"],"datePosted":"2026-04-18T16:00:12.545Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Software development, Distributed systems, Public cloud platforms, Containerization & deployment technologies, Orchestration platforms, NoSQL document databases, Structured databases, Software engineering best practices, CI/CD tooling, Data warehouses, Data pipeline/ETL tools, Scaling products at hyper-growth startups, AI technologies","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":252000,"maxValue":315000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a400e696-2d2"},"title":"Staff Software Engineer, Enterprise GenAI","description":"<p>We&#39;re seeking a strong engineer to join our team and help us build and scale our product in a fast-paced environment. As a Staff Software Engineer, you will own large new areas within our product, working across backend, frontend, and interacting with LLMs and ML models. You will solve hard engineering problems in scalability and reliability.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Delivering experiments at a high velocity and level of quality to engage our customers</li>\n<li>Working across the entire product lifecycle from conceptualization through production</li>\n<li>Being able, and willing, to multi-task and learn new technologies quickly</li>\n</ul>\n<p>Ideally, you&#39;d have:</p>\n<ul>\n<li>7+ years of full-time engineering experience, post-graduation</li>\n<li>Experience scaling products at hyper growth startups</li>\n<li>Experience tinkering with or productizing LLMs, vector databases, and the other latest AI technologies</li>\n<li>Proficient in Python or Javascript/Typescript, and SQL</li>\n<li>Experience with Kubernetes</li>\n<li>Experience with major cloud providers (AWS, Azure, GCP)</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a400e696-2d2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4569678005","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$248,400-$310,500 USD","x-skills-required":["Python","Javascript/Typescript","SQL","Kubernetes","AWS","Azure","GCP"],"x-skills-preferred":["LLMs","vector databases"],"datePosted":"2026-04-18T16:00:11.482Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Javascript/Typescript, SQL, Kubernetes, AWS, Azure, GCP, LLMs, vector databases","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":248400,"maxValue":310500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f7c042aa-57b"},"title":"Program Manager (Homeland Layered Defense), Public Sector","description":"<p>We&#39;re hiring a Program Manager to lead and own execution for Scale&#39;s portfolio of clients charged with a layered defense for the United States.</p>\n<p>This role is ideal for someone who blends program leadership, systems-building obsession, technical fluency , and who thrives in fast-moving, ambiguous, and mission-driven environments.</p>\n<p>As a Program Manager, you will:</p>\n<p>Serve as the contractual and internal program lead across a large-scale public sector engagement, integrating multiple sub-efforts under one umbrella (e.g., data pipelines for computer vision, software deployments for GenAI solutions) Own relationships with senior government stakeholders and ensure that we meet contractual obligations, performance metrics, and reporting requirements Own relationships with subcontractors to ensure successful delivery of the prime contract Lead formal engagements such as contract reviews, customer program syncs, performance assessments, and reporting cycles Lead and supervise a team of 3-4 delivery managers who run day-to-day customer engagement efforts Oversee a technical delivery team that will drive product roadmap alignment that adapts evolving customer needs</p>\n<p>Must haves:</p>\n<p>An active TS/SCI clearance 10+ years professional experience, ideally managing complex technical programs for DoD or national security customers A proven ability to lead large, cross-functional efforts across delivery, contracts, and customer teams A healthy obsession with establishing and driving adoption of programmatic systems and processes; plus the ability to exert influence and manage associated change Familiarity with FAR-based contracts and program structures like IDIQs, task orders, and OTAs Strong business acumen with the ability to manage performance metrics, contract deliverables, and program budgets A track record of high ownership in ambiguous environments Excellent written and verbal communication skills , especially with senior government stakeholders Deep curiosity about agentic AI and its application to homeland layered defense A willingness to travel up to 25% of the time</p>\n<p>We have a diverse team with a variety of skill sets, many have:</p>\n<p>Proficiency in Python, SQL or other programming languages A proven track record in B2B + B2G client facing roles and expanding client relationships Prior experience delivering technical solutions to government customers Domain expertise in a relevant field (e.g. modeling and simulation, joint planning processes, intelligence workflows, etc)</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f7c042aa-57b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4667857005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$232,800-$291,000 USD","x-skills-required":["TS/SCI clearance","10+ years professional experience","program management","technical leadership","business acumen","communication skills","agentic AI","homeland layered defense"],"x-skills-preferred":["Python","SQL","programming languages","B2B + B2G client facing roles","technical solutions delivery","domain expertise"],"datePosted":"2026-04-18T16:00:07.619Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, DC"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"TS/SCI clearance, 10+ years professional experience, program management, technical leadership, business acumen, communication skills, agentic AI, homeland layered defense, Python, SQL, programming languages, B2B + B2G client facing roles, technical solutions delivery, domain expertise","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":232800,"maxValue":291000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bd7327f8-fcf"},"title":"Staff Software Engineer, Full-Stack - Enterprise Gen AI","description":"<p>We&#39;re looking for a frontend-focused full-stack engineer to help build AI-powered applications that redefine enterprise workflows and push the boundaries of interactive AI. As a staff software engineer, you&#39;ll work on a mix of cutting-edge customer-facing AI applications and internal SaaS products. Our engineering team powers projects like TIME&#39;s Person of the Year AI experience, where our AI technology helped shape one of the most iconic features in media. You&#39;ll also contribute to Scale&#39;s GenAI Platform (SGP), a powerful system that enables businesses to build and deploy AI agents at scale.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Building and enhancing user-facing AI applications for major enterprise customers, including high-profile media and Fortune 500 companies</li>\n<li>Developing and refining features for Scale&#39;s GenAI Platform, empowering businesses to build, deploy, and manage AI-driven agents</li>\n<li>Designing, building, and optimizing polished, high-performance UIs using Next.js, React, TypeScript, and Tailwind</li>\n<li>Working closely with product managers, designers, and AI/ML teams to create seamless, intuitive, and impactful user experiences</li>\n<li>Integrating frontend applications with backend services, working with APIs, authentication systems, and cloud-based infrastructure</li>\n</ul>\n<p>In this role, you&#39;ll have the opportunity to shape the future of AI-powered user experiences, working on projects that impact millions of users while developing tools that empower businesses to deploy AI at scale.</p>\n<p>The base salary range for this full-time position in our hub locations of San Francisco, New York, or Seattle is $248,400,$310,500 USD. Compensation packages at Scale include base salary, equity, and benefits. You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bd7327f8-fcf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4529529005","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$248,400—$310,500 USD","x-skills-required":["Next.js","React","TypeScript","Tailwind","AI/ML","APIs","Authentication systems","Cloud-based infrastructure"],"x-skills-preferred":["FastAPI","PostgreSQL","GraphQL","AWS","Azure","GCP","Data-rich web platforms","Interactive AI applications","Agent-based systems"],"datePosted":"2026-04-18T16:00:02.231Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY; San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Next.js, React, TypeScript, Tailwind, AI/ML, APIs, Authentication systems, Cloud-based infrastructure, FastAPI, PostgreSQL, GraphQL, AWS, Azure, GCP, Data-rich web platforms, Interactive AI applications, Agent-based systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":248400,"maxValue":310500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_44975b06-cb1"},"title":"Senior Full-Stack Software Engineer, (Forward Deployed), GPS","description":"<p>We&#39;re seeking a Senior Full-Stack Software Engineer to join our Global Public Sector team. As a forward-deployed engineer, you&#39;ll collaborate directly with public sector counterparts to build full-stack, AI applications that solve critical challenges and achieve meaningful impact for citizens.</p>\n<p>Our core work consists of creating custom AI applications, generating high-quality training data for custom LLMs, and upskilling and advisory services to spread the impact of AI.</p>\n<p>You&#39;ll partner with public sector clients to scope, collect feedback, and implement solutions for complex problems. You&#39;ll also architect production-grade applications that integrate AI models with full-stack frameworks, manage infrastructure within cloud environments, and contribute to core platform features.</p>\n<p>Ideally, you&#39;ll have a Bachelor&#39;s degree in Computer Science or a related quantitative field, 5+ years of full-stack engineering experience, and proficiency in React, TypeScript, Next.js, Python, Node.js, PostgreSQL or MongoDB, and hands-on experience with Docker, Kubernetes, and Azure/AWS/GCP.</p>\n<p>We&#39;re looking for a self-starting approach with technical maturity to navigate ambiguous requirements and deliver reliable software. You&#39;ll also need to drive async communication methodologies to reduce communication frictions.</p>\n<p>If you&#39;re ready to shape the future of AI in the public sector and be a founding member of our team, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_44975b06-cb1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4673310005","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["React","TypeScript","Next.js","Python","Node.js","PostgreSQL","MongoDB","Docker","Kubernetes","Azure","AWS","GCP"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:59.289Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Next.js, Python, Node.js, PostgreSQL, MongoDB, Docker, Kubernetes, Azure, AWS, GCP"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bd00b53a-6fa"},"title":"Software Engineer, Enterprise AI","description":"<p>We are seeking a strong engineer to join our team and help us build and scale our product in a fast-paced environment. The ideal candidate will have a strong understanding of software engineering principles and practices, as well as experience with large-scale distributed systems.</p>\n<p>You will be responsible for owning large new areas within our product, working across backend, frontend, and interacting with LLMs and ML models. You will solve hard engineering problems in scalability and reliability.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Owning large new areas within our product</li>\n<li>Working across backend, frontend, and interacting with LLMs and ML models</li>\n<li>Delivering experiments at a high velocity and level of quality to engage our customers</li>\n<li>Working across the entire product lifecycle from conceptualization through production</li>\n</ul>\n<p>Ideally, you&#39;d have:</p>\n<ul>\n<li>4+ years of full-time engineering experience, post-graduation</li>\n<li>Experience scaling products at hyper growth startups</li>\n<li>Experience tinkering with or productizing LLMs, vector databases, and the other latest AI technologies</li>\n<li>Proficient in Python or Javascript/Typescript, and SQL</li>\n<li>Experience with Kubernetes</li>\n<li>Experience with major cloud providers (AWS, Azure, GCP)</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bd00b53a-6fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4513943005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$179,400-$224,250 USD","x-skills-required":["Python","Javascript/Typescript","SQL","Kubernetes","AWS","Azure","GCP"],"x-skills-preferred":["LLMs","vector databases","AI technologies"],"datePosted":"2026-04-18T15:59:58.329Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY; San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Javascript/Typescript, SQL, Kubernetes, AWS, Azure, GCP, LLMs, vector databases, AI technologies","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":179400,"maxValue":224250,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_45fc6ed2-285"},"title":"Senior Full-Stack Software Engineer, (Forward Deployed), GPS","description":"<p>We&#39;re seeking a Senior Full-Stack Software Engineer to join our Global Public Sector team. As a forward-deployed engineer, you&#39;ll collaborate directly with public sector counterparts to build full-stack AI applications that solve their most pressing challenges.</p>\n<p>Our core work consists of creating custom AI applications, generating high-quality training data for custom LLMs, and upskilling and advisory services to spread the impact of AI.</p>\n<p>You&#39;ll partner with public sector clients to scope, collect feedback, and implement solutions for complex problems. You&#39;ll also architect production-grade applications that integrate AI models with full-stack frameworks, manage infrastructure within cloud environments, and contribute to core platform features.</p>\n<p>Ideally, you&#39;ll have a Bachelor&#39;s degree in Computer Science or a related quantitative field, 5+ years of full-stack engineering experience, and proficiency in React, TypeScript, Next.js, Python, Node.js, PostgreSQL or MongoDB, Docker, Kubernetes, and Azure/AWS/GCP.</p>\n<p>You&#39;ll be a self-starting individual with technical maturity to navigate ambiguous requirements and deliver reliable software. You&#39;ll also have experience working directly within customer infrastructure to deploy, maintain, and troubleshoot complex, end-to-end solutions.</p>\n<p>Nice to have: proficient in Arabic, past experience working in a forward-deployed engineer/dedicated customer engineer role, experience working cross-functionally with operations, and experience building solutions with LLMs and a deep understanding of the overall Gen AI landscape.</p>\n<p>Please note that our policy requires a 90-day waiting period before reconsidering candidates for the same role.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_45fc6ed2-285","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4676606005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["React","TypeScript","Next.js","Python","Node.js","PostgreSQL","MongoDB","Docker","Kubernetes","Azure","AWS","GCP"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:52.395Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Doha, Qatar"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Next.js, Python, Node.js, PostgreSQL, MongoDB, Docker, Kubernetes, Azure, AWS, GCP"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_14499a71-fa9"},"title":"Software Engineer, Enterprise","description":"<p>At Scale AI, we&#39;re pioneering the next era of enterprise AI. As businesses race to harness the power of Generative AI, Scale is at the forefront, delivering cutting-edge solutions that transform workflows, automate complex processes, and drive unparalleled efficiency for the largest enterprises.</p>\n<p>We&#39;re looking for a Backend Engineer to help bring large-scale GenAI systems to production. In this role, you&#39;ll build the core infrastructure that powers AI products for some of the world&#39;s largest enterprises,designing scalable APIs, distributed data systems, and robust deployment pipelines that enable production-grade reliability and performance.</p>\n<p>This is a rare opportunity to be at the center of the GenAI revolution, solving hard backend and infrastructure challenges that make AI truly work at enterprise scale. If you&#39;re excited about shaping how AI systems are deployed and scaled in the real world, we want to hear from you.</p>\n<p>At Scale, we don&#39;t just follow AI advancements , we lead them. Backed by deep expertise in data, infrastructure, and model deployment, we are uniquely positioned to solve the hardest problems in AI adoption. Join us in shaping the future of enterprise AI, where your work will directly impact how businesses operate, innovate, and grow in the age of GenAI.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, build, and scale backend systems that power enterprise GenAI products, focusing on reliability, performance, and deployment across both Scale&#39;s and customers&#39; infrastructure.</li>\n</ul>\n<ul>\n<li>Develop core services and APIs that integrate AI models and enterprise data sources securely and efficiently, enabling production-scale AI adoption.</li>\n</ul>\n<ul>\n<li>Architect scalable distributed systems for data processing, inference, and orchestration of large-scale GenAI workloads.</li>\n</ul>\n<ul>\n<li>Optimize backend performance for latency, throughput, and cost,ensuring AI applications can operate at enterprise scale across hybrid and multi-cloud environments.</li>\n</ul>\n<ul>\n<li>Manage and evolve cloud infrastructure (AWS, Azure, or GCP), driving automation, observability, and security for large-scale AI deployments.</li>\n</ul>\n<ul>\n<li>Collaborate with ML and product teams to bring cutting-edge GenAI models into production through efficient APIs, model serving systems, and evaluation frameworks.</li>\n</ul>\n<ul>\n<li>Continuously improve reliability and scalability, applying strong engineering practices to make AI systems robust, maintainable, and enterprise-ready.</li>\n</ul>\n<p><strong>Ideal Candidate</strong></p>\n<ul>\n<li>4+ years of experience developing large-scale backend or infrastructure systems, with a strong emphasis on distributed services, reliability, and scalability.</li>\n</ul>\n<ul>\n<li>Proficiency in Python or TypeScript, with experience designing high-performance APIs and backend architectures using frameworks such as FastAPI, Flask, Express, or NestJS.</li>\n</ul>\n<ul>\n<li>Deep familiarity with cloud infrastructure (AWS and Azure preferred), including container orchestration (Kubernetes, Docker) and Infrastructure-as-Code tools like Terraform.</li>\n</ul>\n<ul>\n<li>Experience managing data systems such as relational and NoSQL databases (PostgreSQL, DynamoDB, etc.) and building pipelines for data-intensive applications.</li>\n</ul>\n<ul>\n<li>Hands-on experience with GenAI applications, model integration, or AI agent systems,understanding how to deploy, evaluate, and scale AI workloads in production.</li>\n</ul>\n<ul>\n<li>Strong understanding of observability, CI/CD, and security best practices for running services in enterprise or multi-tenant environments.</li>\n</ul>\n<ul>\n<li>Ability to balance rapid iteration with production-grade quality, shipping reliable backend systems in fast-paced environments.</li>\n</ul>\n<p>Collaborative mindset, working closely with ML, infra, and product teams to bring complex GenAI systems into production at enterprise scale.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_14499a71-fa9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4536653005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","TypeScript","FastAPI","Flask","Express","NestJS","AWS","Azure","Kubernetes","Docker","Terraform","PostgreSQL","DynamoDB","GenAI","Model Integration","AI Agent Systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:48.948Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, TypeScript, FastAPI, Flask, Express, NestJS, AWS, Azure, Kubernetes, Docker, Terraform, PostgreSQL, DynamoDB, GenAI, Model Integration, AI Agent Systems"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_78eea632-7b6"},"title":"Deep Research Agent Tech Lead","description":"<p>We&#39;re seeking a highly technical and strategic Staff/Senior Staff Machine Learning Engineer to act as the Tech Lead for our next-generation deep research agents for the Enterprise.</p>\n<p>This high-impact role will drive the technical direction and oversight for Deep Research Agent Development, translating cutting-edge research in Generative AI, Large Language Models (LLMs), and Agentic Frameworks into robust, scalable, and high-impact production systems that enhance enterprise operations, analytics, and core efficiency.</p>\n<p>The ideal candidate thrives in a fast-paced environment, has a passion for both deep technical work and mentoring, and is capable of setting a long-term technical strategy for a critical domain while maintaining a strong, hands-on delivery focus.</p>\n<p><strong>Responsibilities</strong></p>\n<p><strong>Technical Leadership &amp; Vision</strong></p>\n<ul>\n<li>Set the Technical Roadmap: Define and own the technical strategy, architecture, and roadmap for Deep Research Agents for the Enterprise, ensuring alignment with Scale AI’s overall AI strategy and business goals.</li>\n</ul>\n<ul>\n<li>Drive Breakthrough Research to Production: Lead the end-to-end development, from initial research to production deployment, to landing on customer impact, with a focus on integrating diverse data modalities.</li>\n</ul>\n<ul>\n<li>Core Agent Capabilities Development:</li>\n</ul>\n<p><strong>Advanced Knowledge Retrieval</strong>: Architect and implement state-of-the-art retrieval systems to ensure the agents provide accurate and comprehensive answers from public and proprietary data sources from enterprises.</p>\n<p><strong>Data Analysis</strong>: Design and champion the development of data analysis agents that accurately translate complex natural language queries into executable SQL/code against diverse enterprise data schemas.</p>\n<p><strong>Multimodal Intelligence</strong>: Lead the integration of Multimodal AI capabilities to process and extract structured information from visual documents, tables, and forms, enriching the agent&#39;s knowledge base.</p>\n<p><strong>Architecture &amp; Design</strong>: Design and champion highly scalable, reliable, and low-latency infrastructure and frameworks for building, orchestrating, and evaluating multi-agent systems at enterprise scale.</p>\n<p><strong>Technical Excellence</strong>: Serve as the technical authority for the team, leading design reviews, defining ML engineering best practices, and ensuring code quality, security, and operational excellence for all agent systems.</p>\n<p><strong>Team Leadership &amp; Mentorship</strong></p>\n<ul>\n<li>Lead and Mentor: Technically lead and mentor a team of Machine Learning Engineers and Research Scientists, fostering a culture of innovation, rigorous engineering, rapid iteration, and technical depth.</li>\n</ul>\n<ul>\n<li>Recruiting &amp; Growth: Partner with management to hire, onboard, and grow top-tier talent, helping to shape the long-term structure and capabilities of the team.</li>\n</ul>\n<ul>\n<li>Cross-Functional Influence: Collaborate effectively with Product Managers, Data Scientists, and other engineering/science teams to translate ambiguous, high-level business problems into concrete, executable technical specifications and impactful agent solutions.</li>\n</ul>\n<p><strong>Basic Qualifications</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, Electrical Engineering, a related field, or equivalent practical experience.</li>\n</ul>\n<ul>\n<li>8+ years of experience in software development, with at least 6 years focused on Machine Learning, Deep Learning, or Applied Research in a production environment.</li>\n</ul>\n<ul>\n<li>2+ years of experience in a formal or informal Technical Leadership role (Team Lead, Tech Lead) with a focus on setting technical direction for a domain.</li>\n</ul>\n<ul>\n<li>Deep expertise in Generative AI and Large Language Models (LLMs).</li>\n</ul>\n<ul>\n<li>Demonstrated experience designing, building, and deploying AI Agents or complex Agentic systems in production at scale.</li>\n</ul>\n<ul>\n<li>Experience with large-scale distributed systems and real-time data processing.</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Advanced degree (Master&#39;s or Ph.D.) in Computer Science, Machine Learning, or a related quantitative field.</li>\n</ul>\n<ul>\n<li>Demonstrated experience designing and deploying production-grade Text-to-SQL systems, including handling complex schema linking and query optimization.</li>\n</ul>\n<ul>\n<li>Practical experience with Multimodal AI, specifically integrating OCR and vision-language models for document intelligence and structured data extraction from images/forms.</li>\n</ul>\n<ul>\n<li>Proven experience in one or more relevant deep research areas: Reinforcement Learning (RL), Reasoning and Planning, Agentic Systems.</li>\n</ul>\n<ul>\n<li>Experience with vector databases and advanced retrieval techniques.</li>\n</ul>\n<ul>\n<li>A track record of publishing research papers in top-tier ML/AI conferences (e.g., NeurIPS, ICML, ICLR, KDD).</li>\n</ul>\n<ul>\n<li>Excellent written and verbal communication skills, with the ability to articulate complex technical vision to executive stakeholders and technical peers.</li>\n</ul>\n<ul>\n<li>Experience driving cross-team technical initiatives that have delivered significant business impact.</li>\n</ul>\n<p><strong>Compensation</strong></p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity-based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p><strong>About Us</strong></p>\n<p>At Scale, our mission is to develop reliable AI systems for the world&#39;s most important decisions. Our products provide the high-quality data and full-stack technologies that power the world&#39;s leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact. We work closely with industry leaders like Meta, Cisco, DLA Piper, Mayo Clinic, Time Inc., the Government of Qatar, and U.S. government agencies including the Army and Air Force. We are expanding our team to accelerate the development of AI applications.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_78eea632-7b6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4623590005","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$264,800-$331,000 USD","x-skills-required":["Generative AI","Large Language Models (LLMs)","Agentic Frameworks","Machine Learning","Deep Learning","Applied Research","Distributed Systems","Real-time Data Processing"],"x-skills-preferred":["Text-to-SQL Systems","Multimodal AI","Reinforcement Learning (RL)","Reasoning and Planning","Agentic Systems","Vector Databases","Advanced Retrieval Techniques"],"datePosted":"2026-04-18T15:59:45.270Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Generative AI, Large Language Models (LLMs), Agentic Frameworks, Machine Learning, Deep Learning, Applied Research, Distributed Systems, Real-time Data Processing, Text-to-SQL Systems, Multimodal AI, Reinforcement Learning (RL), Reasoning and Planning, Agentic Systems, Vector Databases, Advanced Retrieval Techniques","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":264800,"maxValue":331000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6365e7d7-511"},"title":"Senior Forward Deployed Data Scientist/Engineer","description":"<p>We&#39;re hiring a Senior Forward Deployed Data Scientist / Engineer to work directly with customers on ambiguous, high-impact problems at the intersection of data science, product development, and AI deployment.</p>\n<p>This is not a traditional analytics role. On this team, data scientists do the core statistical and modeling work, but they also build real tools and products: evaluation explorers, operator workflows, decision-support systems, experimentation surfaces, and customer-specific AI/data applications that get used in production.</p>\n<p>The right candidate is strong in first-principles problem solving, rigorous measurement, and technical execution. They know how to define metrics, design experiments, diagnose failures, and build systems that people actually use. They are also comfortable using modern AI-assisted development tools to prototype and iterate quickly without sacrificing reliability, observability, or judgment. Python and SQL matter in this role, but as execution fluency in service of building better products and making better decisions.</p>\n<p>Responsibilities: Partner directly with enterprise customers to understand workflows, operational pain points, constraints, and success criteria Turn ambiguous business and product problems into measurable solutions with clear metrics, technical designs, and deployment plans Design and build internal and customer-facing data products, including evaluation tools, workflow applications, decision-support systems, and thin product layers on top of data/ML systems Build end-to-end solutions across data ingestion, transformation, experimentation, statistical modeling, deployment, monitoring, and iteration Design evaluation frameworks, benchmarks, and feedback loops for ML/LLM systems, human-in-the-loop workflows, and model-assisted operations Apply rigorous statistical thinking to experimentation, causal inference, metric design, forecasting, segmentation, diagnostics, and performance measurement Use AI-assisted development workflows to accelerate prototyping and product iteration, while maintaining strong engineering discipline Diagnose failure modes across data quality, model behavior, retrieval, workflow design, and user experience, and drive fixes into production Act as the voice of the customer to Product, Engineering, and Data Science, using field learnings to shape roadmap and platform capabilities</p>\n<p>Requirements: 5+ years of experience in data science, machine learning, quantitative engineering, or another highly analytical technical role Proven track record of shipping data, ML, or AI systems that delivered measurable business or product impact Exceptional ability to structure ambiguous problems, define the right success metrics, and translate them into executable technical plans Strong foundation in statistics, experimentation, causal reasoning, and measurement Experience building tools or products, not just analyses , for example internal workflow tools, evaluation systems, operator-facing products, experimentation platforms, or customer-specific applications Hands-on fluency in Python, SQL, and modern data/AI tooling; able to inspect data, prototype quickly, debug deeply, and productionize solutions that work Comfort using AI-assisted coding and development workflows to move from idea to usable product quickly Strong communication and stakeholder management skills; able to work effectively with customers, engineers, product teams, and executives High ownership and bias toward shipping in fast-moving environments with incomplete information</p>\n<p>Preferred qualifications: Experience in a forward deployed, solutions, consulting, or other client-facing technical role Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign</p>\n<p>What success looks like: Success in this role means taking a messy, high-stakes customer problem and turning it into a deployed system that is actually used. Sometimes that system is a model. Sometimes it is an evaluation framework. Sometimes it is an operator-facing tool or a lightweight data product that changes how decisions get made. In all cases, success is defined by measurable impact, rigorous evaluation, and reliable execution.</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p>Salary Range: $167,200-$209,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6365e7d7-511","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4636227005","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$167,200-$209,000 USD","x-skills-required":["Python","SQL","Modern data/AI tooling","Statistics","Experimentation","Causal reasoning","Measurement","Data science","Machine learning","Quantitative engineering"],"x-skills-preferred":["Experience in a forward deployed, solutions, consulting, or other client-facing technical role","Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products","Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow","Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery","Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems","Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling","Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign"],"datePosted":"2026-04-18T15:59:44.618Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Modern data/AI tooling, Statistics, Experimentation, Causal reasoning, Measurement, Data science, Machine learning, Quantitative engineering, Experience in a forward deployed, solutions, consulting, or other client-facing technical role, Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products, Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow, Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery, Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems, Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling, Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":167200,"maxValue":209000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b68ff4cc-e74"},"title":"Data Engineer, Safeguards","description":"<p><strong>About the role</strong></p>\n<p>Anthropic is looking for a Data Engineer to join the Safeguards team and build the data foundations that keep our AI systems safe. The Safeguards team works to monitor models, prevent misuse, and ensure user well-being.</p>\n<p>You&#39;ll design and build the data pipelines, warehousing solutions, and analytical tooling that power our safety and trust efforts at scale. You&#39;ll work closely with engineers, data scientists, and policy teams to ensure the Safeguards organization has the data it needs to detect abuse patterns, measure the effectiveness of safety interventions, and make informed decisions about model behavior and enforcement.</p>\n<p>This is a high-impact role where your work will directly support Anthropic&#39;s mission to develop AI that is safe and beneficial.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, build, and maintain scalable data pipelines that support safety monitoring, abuse detection, and enforcement workflows</li>\n<li>Develop and optimize data models and warehousing solutions to enable efficient analysis of large-scale usage and safety data</li>\n<li>Build and maintain dashboards and reporting infrastructure that give Safeguards teams visibility into model behavior, misuse patterns, and enforcement outcomes</li>\n<li>Collaborate with engineers to integrate data from multiple sources , including model outputs, user reports, and automated classifiers , into a unified analytical layer</li>\n<li>Implement data quality frameworks, monitoring, and alerting to ensure the reliability of safety-critical data</li>\n<li>Partner with research teams to surface data insights that inform model improvements and safety interventions</li>\n<li>Develop self-service data tooling that enables stakeholders to explore safety data and generate reports independently</li>\n<li>Contribute to data governance practices, including access controls, retention policies, and privacy-compliant data handling</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have 3+ years of experience in data engineering, analytics engineering, or a related role</li>\n<li>Are proficient in SQL and Python, with experience building and maintaining ETL/ELT pipelines</li>\n<li>Have hands-on experience with modern data stack tools such as dbt, Airflow, Spark, or similar orchestration and transformation frameworks</li>\n<li>Have worked with cloud data platforms (BigQuery, Redshift, Snowflake, or similar)</li>\n<li>Are comfortable building dashboards and data visualizations using tools like Looker, Tableau, or Metabase</li>\n<li>Communicate clearly and can translate complex data concepts for both technical and non-technical audiences</li>\n<li>Are results-oriented, flexible, and willing to pick up slack even when it falls outside your job description</li>\n<li>Care about the societal impacts of AI and are motivated by safety work</li>\n</ul>\n<p><strong>Strong candidates may have:</strong></p>\n<ul>\n<li>Experience with trust &amp; safety, integrity, fraud, or abuse detection data systems</li>\n<li>Experience with large-scale event streaming systems (Kafka, Pub/Sub, Kinesis)</li>\n<li>Built data infrastructure that supports ML model monitoring or evaluation</li>\n<li>A background in statistical analysis, or experience collaborating closely with data scientists</li>\n<li>Developed internal tooling or self-service analytics platforms</li>\n</ul>\n<p><strong>Strong candidates need not have:</strong></p>\n<ul>\n<li>A formal degree in Computer Science or a related field , we value practical experience and demonstrated ability over credentials</li>\n<li>Prior experience in AI or machine learning , you&#39;ll learn the domain-specific context on the job</li>\n<li>Previous experience at an AI safety or research organization</li>\n<li>Deep expertise across every tool listed above , familiarity with a subset and a willingness to learn is enough</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices. Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills. The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b68ff4cc-e74","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5156057008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"£170,000-£220,000 GBP","x-skills-required":["SQL","Python","ETL/ELT pipelines","dbt","Airflow","Spark","cloud data platforms","BigQuery","Redshift","Snowflake","Looker","Tableau","Metabase"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:33.960Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, ETL/ELT pipelines, dbt, Airflow, Spark, cloud data platforms, BigQuery, Redshift, Snowflake, Looker, Tableau, Metabase","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":170000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_83b76cad-0da"},"title":"Operations Specialist","description":"<p>As an early member of our operations team, you&#39;ll drive revenue by ensuring Scale AI meets customer commitments in a timely manner while maintaining the highest quality standards.</p>\n<p>You will manage our supply operation funnel by building and running solutions, tools, and processes by working with a cross-functional team including Customer Operations, Product Operations, Product Managers, and many others.</p>\n<p>You will be solving problems no one has solved before, and you will need to be relentless in driving stellar results, running pilots, tests, and experiments.</p>\n<p>The blend of operations and ownership of our most important outcomes make this a unique and exciting role at the heart of Scale’s daily operations.</p>\n<p>The ideal candidate is scrappy, analytical, empathetic, outcome focused, and above all someone who drives and inspires results.</p>\n<p>You will:</p>\n<ul>\n<li>Build and drive some of our most critical operational processes</li>\n</ul>\n<ul>\n<li>Own the day-to-day delivery of customer commitments</li>\n</ul>\n<ul>\n<li>Create an effective feedback loop between the front line, product, strategy, and customers</li>\n</ul>\n<ul>\n<li>Collaborate with stakeholders to improve processes for new and existing customers</li>\n</ul>\n<p>Ideally you’d have:</p>\n<ul>\n<li>Advanced English skills</li>\n</ul>\n<ul>\n<li>At least three months of industry experience in an operational role and/or a top-tier consulting firm</li>\n</ul>\n<ul>\n<li>An undergraduate degree with an analytics heavy major (e.g., Engineering or Economics) and/or a graduate degree in Engineering, Economics, or Business</li>\n</ul>\n<ul>\n<li>An action-oriented mindset that balances creative problem solving with the scrappiness to ultimately deliver results</li>\n</ul>\n<ul>\n<li>Analytical, planning, and process improvement capability</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience with reading SQL, or have demonstrated analytical skills</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_83b76cad-0da","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4594237005","x-work-arrangement":null,"x-experience-level":"entry","x-job-type":"contract","x-salary-range":"$11,400 USD annually","x-skills-required":["English","Operational role","Top-tier consulting firm","Analytics heavy major","Graduate degree in Engineering, Economics, or Business","Action-oriented mindset","Analytical, planning, and process improvement capability"],"x-skills-preferred":["SQL"],"datePosted":"2026-04-18T15:59:32.608Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mexico City, MX"}},"employmentType":"CONTRACTOR","occupationalCategory":"Operations","industry":"Technology","skills":"English, Operational role, Top-tier consulting firm, Analytics heavy major, Graduate degree in Engineering, Economics, or Business, Action-oriented mindset, Analytical, planning, and process improvement capability, SQL","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":11400,"maxValue":11400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6cae1ee9-b93"},"title":"Senior Technical Solutions Engineer (Platform)","description":"<p>As a Senior Technical Solutions Engineer, you will provide technical support for Databricks Platform related issues and resolve any challenges involving the Databricks unified analytics platform.</p>\n<p>You will assist customers in their Databricks journey and provide them with the guidance and knowledge that they need to accomplish value and achieve their strategic goals using our products.</p>\n<p>They will look to you for answers to everything from basic technical questions to complex architectural scenarios spanning across the entire Big Data ecosystem.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Troubleshoot and resolve complex customer issues related to Databricks platform</li>\n<li>Provide best practices support for custom-built solutions developed by Databricks customers</li>\n<li>Deliver suggestions for improving performance in customer-specific environments</li>\n<li>Assist with issues around third-party integrations with Databricks environment</li>\n<li>Demonstrate and coordinate with engineering and escalation teams to achieve resolution of customer issues and requests</li>\n<li>Participate in the creation and maintenance of company documentation and knowledge articles</li>\n<li>Be a true proponent of customer advocacy</li>\n<li>Strengthen your AWS/Azure and Databricks platform expertise through learning and internal training programs</li>\n<li>Participate in weekend and weekday on call rotation</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>4+ years experience designing, building, testing, and maintaining Python/Java/Scala based applications</li>\n<li>Expert level knowledge in python is desired</li>\n<li>Strong experience with SQL-based database is required</li>\n<li>Linux/Unix administration skills</li>\n<li>Hands-on experience with AWS, Azure or GCP</li>\n<li>Experience with &quot;Distributed Big Data Computing&quot; environment</li>\n<li>Technical degree or the equivalent experience</li>\n<li>Written and spoken proficiency in both Japanese and English</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6cae1ee9-b93","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8488552002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","Scala","SQL","Linux/Unix","AWS","Azure","GCP","Distributed Big Data Computing"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:28.244Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Tokyo, Japan"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, SQL, Linux/Unix, AWS, Azure, GCP, Distributed Big Data Computing"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5465ae79-24e"},"title":"Analytics & Systems Lead, Finance","description":"<p>We&#39;re seeking an experienced Analytics &amp; Systems Lead to join our Finance team. In this role, you will work closely with stakeholders across business finance, corporate finance, and accounting to design and develop internal tools and AI agents that automate workflows across finance and accounting.</p>\n<p>You will design data models, prototype internal tools, and implement agent-driven workflows using our internal data infrastructure, system integration tooling, Scale&#39;s proprietary AI platform, and emerging AI tools.</p>\n<p>Key responsibilities include designing and developing end-to-end agent-driven workflows, building scalable data models and pipelines, partnering with stakeholders to translate business requirements into technical requirements, and collaborating with engineering teams to develop internal tools.</p>\n<p>Ideal candidates will have 5+ years of experience in data analytics, analytics engineering, or data science roles, expert knowledge of SQL and Python, and experience building internal tools, automation systems, or data products.</p>\n<p>Compensation packages at Scale include base salary, equity, and benefits, with a base salary range of $200,000-$250,000 USD for this full-time position in San Francisco.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5465ae79-24e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4673090005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$200,000-$250,000 USD","x-skills-required":["SQL","Python","data analysis","data modeling","internal tools","automation systems","data products"],"x-skills-preferred":["JavaScript","modern automation tooling"],"datePosted":"2026-04-18T15:59:26.227Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, data analysis, data modeling, internal tools, automation systems, data products, JavaScript, modern automation tooling","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":200000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_130450d3-e02"},"title":"Analytics Lead, GenAI Marketplace","description":"<p>We&#39;re looking for an experienced Senior Data Analyst to join our Growth Operations team. As a key member of our team, you will partner with leadership, operators, and engineers to help unpack ambiguous problems and build lasting, scalable analytics solutions.</p>\n<p>Your work will directly evolve how we operate and measure our growth strategies, funnels, and pipelines. You will be responsible for building scalable data assets, adding contextual layers to data, and providing insights and conclusions to stakeholders.</p>\n<p>To succeed in this role, you will need to be detail-oriented, rigorous about validating results, and talented at distilling down complexity. You will also need to love tackling and solving hard problems.</p>\n<p>In addition to building scalable data assets, you will also be responsible for partnering with Data Engineers, Data Scientists, and cross-functional stakeholders to develop business metrics to understand our current performance and influence our roadmaps.</p>\n<p>If you have 3+ years of industry experience in a highly analytical role, a degree in a quantitative field, and expert-level proficiency in writing complex SQL queries, we encourage you to apply.</p>\n<p>As a Senior Data Analyst at Scale, you will have the opportunity to work with a talented team of professionals and contribute to the development of reliable AI systems for the world&#39;s most important decisions.</p>\n<p>The ideal candidate will have excellent communication and presentation skills to executives, expertise at defining the right metrics, and diagnosing and understanding data inconsistencies.</p>\n<p>Compensation packages at Scale include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors.</p>\n<p>You&#39;ll also receive benefits including comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_130450d3-e02","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4631695005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$184,000-$230,000 USD","x-skills-required":["SQL","Tableau","Python","Data analysis","Data visualization"],"x-skills-preferred":["Machine learning","Deep learning","Cloud computing"],"datePosted":"2026-04-18T15:59:17.724Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Tableau, Python, Data analysis, Data visualization, Machine learning, Deep learning, Cloud computing","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":184000,"maxValue":230000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_150ca1e8-f29"},"title":"Research Economist, Economic Research","description":"<p>As a Research Economist at Anthropic, you will work to measure and understand AI&#39;s effects on the global economy. You will make fundamental contributions to the development of the Anthropic Economic Index, establishing new methodologies to measure the usage, diffusion, and impact of AI throughout the economy using privacy-preserving tools and novel data sources. You will use frontier methods in econometrics, machine learning, and structural estimation. Such rigour will drive impact, shaping both policy discussions externally and informing Anthropic’s internal business and product decisions.</p>\n<p>Our team combines rigorous empirical methods with novel measurement approaches. We&#39;re building first-of-its-kind datasets tracking AI&#39;s impact on labor markets, productivity, and economic transformation. Using our privacy-preserving measurement system (Clio), we analyze millions of real-world AI interactions to understand how AI augments and automates work across different occupations and tasks.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Make fundamental contributions to the development and expansion of the Anthropic Economic Index, including quarterly reports and industry-specific deep dives</li>\n</ul>\n<ul>\n<li>Design and conduct empirical research on AI&#39;s economic effects, drawing on external data sources and the privacy-preserving measurement systems internally</li>\n</ul>\n<ul>\n<li>Develop new methodological approaches for studying AI&#39;s impact on:</li>\n</ul>\n<ul>\n<li>Labor markets and the future of work</li>\n</ul>\n<ul>\n<li>Productivity and task transformation</li>\n</ul>\n<ul>\n<li>Economic inequality and displacement</li>\n</ul>\n<ul>\n<li>Industry-specific disruption and adaptation</li>\n</ul>\n<ul>\n<li>Aggregate economic trajectories (GDP, productivity, unemployment) under varying AI-adoption scenarios</li>\n</ul>\n<ul>\n<li>Develop causal-inference tooling , e.g. surrogate indexes, heterogeneous-effect pipelines , to help Anthropic evaluate the downstream economic consequences of its own compute, product, and pricing decisions</li>\n</ul>\n<ul>\n<li>Build and maintain relationships with academic institutions, policy think tanks, and other research partners</li>\n</ul>\n<ul>\n<li>Work cross-functionally with other technical teams to improve our measurement infrastructure and data collection</li>\n</ul>\n<ul>\n<li>Translate research insights into actionable recommendations for both product decisions and policy discussions</li>\n</ul>\n<ul>\n<li>Amplify external engagement through research publications, policy briefs, and presentations to diverse stakeholders</li>\n</ul>\n<p>You May Be a Good Fit If You Have:</p>\n<ul>\n<li>PhD in Economics</li>\n</ul>\n<ul>\n<li>Strong track record of empirical research, particularly studies combining novel data sources and economic theory or those implementing frontier methods in causal inference and machine learning</li>\n</ul>\n<ul>\n<li>Experience relevant to the study of AI’s impact on the economy, including:</li>\n</ul>\n<ul>\n<li>Labor market analysis and occupational change</li>\n</ul>\n<ul>\n<li>Task-based approaches to technological transformation</li>\n</ul>\n<ul>\n<li>Large-scale data analysis and econometric methods</li>\n</ul>\n<ul>\n<li>Large language models for social science research</li>\n</ul>\n<ul>\n<li>Policy-relevant economic research</li>\n</ul>\n<ul>\n<li>Experimental and quasi-experimental methods for causal inference</li>\n</ul>\n<ul>\n<li>Macroeconomic modeling and time series forecasting</li>\n</ul>\n<ul>\n<li>Agent-based modeling or large-scale simulation</li>\n</ul>\n<ul>\n<li>Technical skills including:</li>\n</ul>\n<ul>\n<li>Proficiency in Python, R, SQL, or similar tools for large-scale data analysis</li>\n</ul>\n<ul>\n<li>Experience working with novel datasets and measurement systems</li>\n</ul>\n<ul>\n<li>Comfort learning new technical tools and frameworks</li>\n</ul>\n<ul>\n<li>Demonstrated ability to:</li>\n</ul>\n<ul>\n<li>Lead complex research projects from conception to publication</li>\n</ul>\n<ul>\n<li>Communicate technical findings to diverse audiences</li>\n</ul>\n<ul>\n<li>Build relationships across academic, policy, and industry communities</li>\n</ul>\n<ul>\n<li>Strong interest in ensuring AI development benefits humanity</li>\n</ul>\n<ul>\n<li>Comfort working with AI systems and ability to think critically about their capabilities and limitations</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_150ca1e8-f29","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5018472008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$300,000-$405,000 USD","x-skills-required":["PhD in Economics","Empirical research","Econometrics","Machine learning","Structural estimation","Python","R","SQL","Large-scale data analysis","Novel datasets and measurement systems","Causal inference","Macroeconomic modeling","Time series forecasting","Agent-based modeling"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:14.877Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"PhD in Economics, Empirical research, Econometrics, Machine learning, Structural estimation, Python, R, SQL, Large-scale data analysis, Novel datasets and measurement systems, Causal inference, Macroeconomic modeling, Time series forecasting, Agent-based modeling","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":300000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_057b8651-835"},"title":"AI Strategy Consultant, Frontier Tech","description":"<p>As a member of our Frontier Tech Consultant team, you will play a critical role in advancing cutting-edge AI innovations by conducting high-impact experiments and ensuring seamless execution at the highest quality standards.</p>\n<p>Your work will directly contribute to Scale AI’s growth, shaping the future of artificial intelligence. In this role, you will be working on various types of projects, including but not limited to: research experiments, dataset generation, data quality improvements, and in-depth technical analysis.</p>\n<p>You will tackle complex, technical and operational challenges while collaborating closely with Scale’s ML research scientists and SPM team.</p>\n<p>The ideal candidate is analytical, detail-oriented, and results-driven, with strong problem-solving abilities and excellent communication skills.</p>\n<p>We are looking for someone who thrives in a fast-paced environment, is proactive in overcoming challenges, and is committed to delivering exceptional outcomes.</p>\n<p>If you are eager to contribute to the forefront of AI innovation, we encourage you to apply.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and execute research experiments</li>\n<li>Build and evaluate frontier LLM datasets</li>\n<li>Develop training and testing material for frontier pipelines</li>\n<li>Improve quality of existing and new products</li>\n</ul>\n<p>Ideally you’d have:</p>\n<ul>\n<li>Strong machine learning knowledge, either by being in the final years of a ML PhD career or having already graduated</li>\n<li>Strong writing and verbal communication skills</li>\n<li>An action-oriented mindset that balances creative problem solving with the scrappiness to ultimately deliver results</li>\n<li>Analytical, planning, and process improvement capability</li>\n<li>Experience working in a fast-paced, entrepreneurial environment</li>\n<li>Technical skills including familiarity with Python, GPU, AWS, API, LLM, ML, and SQL</li>\n</ul>\n<p>Pay: $60-80/hr</p>\n<p>Commitment: This is a fully remote, US-based part-time (10-20 hours per week), on-going contract position staffed via HireArt.</p>\n<p>HireArt values diversity and is an Equal Opportunity Employer. We are interested in every qualified candidate who is eligible to work in the United States. Unfortunately, we are not able to sponsor visas, including CPT/OPT or employ corp-to-corp.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_057b8651-835","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4472223005","x-work-arrangement":"remote","x-experience-level":null,"x-job-type":"contract","x-salary-range":"$60-80/hr","x-skills-required":["Python","GPU","AWS","API","LLM","ML","SQL"],"x-skills-preferred":["Machine Learning","Data Analysis","Problem Solving"],"datePosted":"2026-04-18T15:59:01.983Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"jobLocationType":"TELECOMMUTE","employmentType":"CONTRACTOR","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, GPU, AWS, API, LLM, ML, SQL, Machine Learning, Data Analysis, Problem Solving"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_28d86251-85e"},"title":"Senior Data Analyst, Marketing Analytics","description":"<p>As a Senior Data Analyst, Marketing Analytics at GitLab, you&#39;ll serve as a strategic analytics partner to Marketing leadership and help shape how we measure, model, and improve marketing performance.</p>\n<p>Reporting to the Senior Manager of Marketing Analytics, you&#39;ll take ownership of high-impact analytical work across attribution, inbound funnel analysis, target setting, and campaign measurement.</p>\n<p>You&#39;ll work across the full data lifecycle, from building and maintaining dbt models in Snowflake that power business intelligence, to delivering executive-ready insights, to partnering cross-functionally to support important business decisions.</p>\n<p>In this role, you&#39;ll use a modern analytics and AI-enabled workflow that includes Snowflake, Claude with MCP connections, and GitLab Duo Agent Platform in git-based team workflows.</p>\n<p>This is a high-visibility role for someone with strong SQL skills, sound judgment in B2B software marketing analytics, and the ability to turn complex data into clear stories for senior stakeholders in our all-remote, values-driven environment.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Serve as a trusted advisor to senior Marketing and Developer Relations leaders, providing data-driven insights that inform marketing strategy, product development, and go-to-market decisions.</li>\n</ul>\n<ul>\n<li>Establish and scale proactive insights across the Marketing function, including experimentation and connecting marketing credit from initial touchpoints to annual recurring revenue.</li>\n</ul>\n<ul>\n<li>Measure the impact of community programs, developer advocacy, and evangelism efforts on awareness, adoption, and pipeline.</li>\n</ul>\n<ul>\n<li>Evaluate the effectiveness of demand generation programs, including trial conversion analysis across segments, lead scoring optimization, and outreach platform performance.</li>\n</ul>\n<ul>\n<li>Partner closely with Analytics Engineering and Data Platform teams to translate business questions into clear technical requirements and support reliable delivery in the analytics stack.</li>\n</ul>\n<ul>\n<li>Help evolve GitLab&#39;s multi-touch attribution approach, including attributed metric definitions and implementation across reporting.</li>\n</ul>\n<ul>\n<li>Contribute to annual planning cycles through partnership with Sales, Finance, and Product leadership.</li>\n</ul>\n<ul>\n<li>Use AI-enabled analytics workflows through GitLab Duo Agent Platform, Claude with MCP integrations, and Glean within GitLab to improve how insights are developed and shared.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>5+ years of experience in a data analyst or analytics engineer role, preferably in B2B software marketing analytics or in a large-scale analytics consultancy supporting similar companies.</li>\n</ul>\n<ul>\n<li>Strong proficiency in SQL, including the ability to connect disparate data sources and build a clear view of marketing program performance in a cloud data warehouse, ideally Snowflake.</li>\n</ul>\n<ul>\n<li>Experience working with senior leaders and tailoring data-driven narratives to support decision-making.</li>\n</ul>\n<ul>\n<li>Hands-on experience with dbt for data transformation, testing, and documentation.</li>\n</ul>\n<ul>\n<li>Strong understanding of B2B marketing funnel metrics, including marketing qualified leads, sales accepted opportunities, pipeline, multi-touch attribution, and conversion rate analysis.</li>\n</ul>\n<ul>\n<li>Experience working with marketing automation platforms such as Marketo and customer relationship management systems such as Salesforce, including data integration patterns and sync logic.</li>\n</ul>\n<ul>\n<li>Proficiency with a business intelligence and visualization tool, with Tableau preferred.</li>\n</ul>\n<ul>\n<li>Excellent written and verbal communication skills, with the ability to distill complex analysis into clear narratives for technical and non-technical audiences.</li>\n</ul>\n<ul>\n<li>Experience with AI-assisted analytics development, including MCP connections and agentic workflows.</li>\n</ul>\n<ul>\n<li>Familiarity with Git workflows for version control, continuous integration and continuous delivery, issue tracking, and merge request workflows.</li>\n</ul>\n<ul>\n<li>Comfort working in a remote, async-first, globally distributed environment.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>The Marketing Analytics team sits within GitLab&#39;s Enterprise Data organization and partners closely with the broader Marketing team as well as other departments across GitLab.</p>\n<p>We build and maintain trusted, scalable data products that help inform strategic decisions for Marketing while also creating alignment with related business metrics, especially across Product and Sales.</p>\n<p>You&#39;ll join a fully remote team that collaborates asynchronously across time zones using our shared standards, code review, and documentation to support consistency and quality.</p>\n<p>We work on analytically complex, high-impact problems where strong data foundations, clear definitions, and thoughtful cross-functional partnerships are essential to helping stakeholders understand performance and make better decisions.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Benefits to support your health, finances, and well-being</li>\n</ul>\n<ul>\n<li>Flexible Paid Time Off</li>\n</ul>\n<ul>\n<li>Team Member Resource Groups</li>\n</ul>\n<ul>\n<li>Equity Compensation &amp; Employee Stock Purchase Plan</li>\n</ul>\n<ul>\n<li>Growth and Development Fund</li>\n</ul>\n<ul>\n<li>Parental leave</li>\n</ul>\n<ul>\n<li>Home office support</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_28d86251-85e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"GitLab","sameAs":"https://about.gitlab.com/","logo":"https://logos.yubhub.co/about.gitlab.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gitlab/jobs/8472178002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$94,100-$201,600 USD","x-skills-required":["SQL","dbt","Snowflake","Tableau","Marketo","Salesforce","AI-assisted analytics development","MCP connections","agentic workflows","Git workflows","version control","continuous integration and continuous delivery","issue tracking","merge request workflows"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:57.489Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, North America"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, dbt, Snowflake, Tableau, Marketo, Salesforce, AI-assisted analytics development, MCP connections, agentic workflows, Git workflows, version control, continuous integration and continuous delivery, issue tracking, merge request workflows","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":94100,"maxValue":201600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e6c2906a-625"},"title":"Senior Software Engineer,  Full-Stack – Scale GP","description":"<p>We are seeking a strong Senior Full-Stack Engineer to help us build, scale, and refine our rapidly growing Generative AI platform, Scale GP. As a senior engineer, you will work across the stack,from React/TypeScript frontends to Python-based backends,while integrating with LLMs and machine learning systems. You will solve complex challenges in scalability, reliability, and product experience while owning significant product areas in a fast-paced environment.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Own major full-stack product areas, driving features from design through production deployment.</li>\n<li>Build modern frontend experiences using React and TypeScript, ensuring performance, usability, and responsiveness.</li>\n<li>Develop reliable backend services in Python, working with distributed systems, data pipelines, and ML/LLM components.</li>\n<li>Integrate with LLMs, vector databases, and AI infrastructure to power intelligent product experiences.</li>\n<li>Deliver experiments and new features quickly, maintaining high quality and tight feedback loops with customers.</li>\n<li>Collaborate across product, ML, and infrastructure teams to shape the direction of Scale GP.</li>\n<li>Adapt quickly,learning new technologies, frameworks, and tools as needed across the stack.</li>\n</ul>\n<p><strong>Ideal Experience</strong></p>\n<ul>\n<li>5+ years of full-time engineering experience, post-graduation.</li>\n<li>Strong experience developing full-stack applications using React, TypeScript, and Python.</li>\n<li>Experience scaling or shipping products at high-growth startups.</li>\n<li>Familiarity with LLMs, vector databases, embeddings, or other modern AI tooling (tinkering or production experience welcome).</li>\n<li>Proficiency with SQL and modern API development.</li>\n<li>Experience with Kubernetes, containerization, and microservice architectures.</li>\n<li>Experience working with at least one major cloud provider (AWS, GCP, or Azure).</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e6c2906a-625","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4637484005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["React","TypeScript","Python","LLMs","vector databases","embeddings","SQL","API development","Kubernetes","containerization","microservice architectures","cloud providers (AWS, GCP, or Azure)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:56.168Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Python, LLMs, vector databases, embeddings, SQL, API development, Kubernetes, containerization, microservice architectures, cloud providers (AWS, GCP, or Azure)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_714b56e4-2cb"},"title":"Strategic Projects Lead - Coding","description":"<p>We&#39;re looking for a Strategic Projects Lead to lead initiatives that will drive $XXM+ in new revenue for our Generative AI business unit. As a key member of our Gen AI Operations team, you will be responsible for overseeing customer pipelines, designing data taxonomy, building pipeline infrastructure, training and coaching teams, and partnering with stakeholders to drive advancements for the largest LLMs in the world.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Serving as the full owner of our most visible and high impact customer pipelines, making decisions that directly impact data quality, operational efficiency, revenue, and margins</li>\n<li>Understanding customer requirements and designing data taxonomy best suited to improving model performance based on customer needs</li>\n<li>Building out pipeline infrastructure to ensure quality and efficiency</li>\n<li>Training, coaching, and managing dynamic and global teams</li>\n<li>Building analytics to make data-driven decisions</li>\n<li>Partnering with diverse stakeholders (Engineering + Product + Ops + Go-to-Market) to work on problems that will drive advancements for the largest LLMs in the world</li>\n</ul>\n<p>You will have:</p>\n<ul>\n<li>A strong technical background, with a degree in Machine Learning Engineering, Computer Science, or Software Engineering</li>\n<li>3+ years of experience leading a team/projects, managing operational processes, or 3+ years of experience as a SWE</li>\n<li>Strong problem-solving capabilities in technical environments, with the ability to come up with creative solutions to complex, ambiguous, operational, and technical problems</li>\n<li>Entrepreneurial experience and mindset, with a passion for building things from scratch and identifying issues and executing quickly</li>\n</ul>\n<p>Please note that our policy requires a 90-day waiting period before reconsidering candidates for the same role.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_714b56e4-2cb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4664232005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Machine Learning Engineering","Computer Science","Software Engineering","SQL","Python","Data Analytics"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:53.295Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mexico City, MX"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning Engineering, Computer Science, Software Engineering, SQL, Python, Data Analytics"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f2bc1be2-478"},"title":"Senior Technical Solutions Engineer, Platform","description":"<p>As a Senior Technical Solutions Engineer, you will provide technical support for Databricks Platform related issues and resolve any challenges involving the Databricks unified analytics platform.</p>\n<p>You will assist customers in their Databricks journey and provide them with the guidance and knowledge that they need to accomplish value and achieve their strategic goals using our products.</p>\n<p>They will look to you for answers to everything from basic technical questions to complex architectural scenarios spanning across the entire Big Data ecosystem.</p>\n<p>You will report to the Senior Manager of Technical Solutions.</p>\n<p>Key responsibilities include: Troubleshooting and resolving complex customer issues related to Databricks platform Providing best practices support for custom-built solutions developed by Databricks customers Delivering suggestions for improving performance in customer-specific environments Assisting with issues around third-party integrations with Databricks environment Demonstrating and coordinating with engineering and escalation teams to achieve resolution of customer issues and requests Participating in the creation and maintenance of company documentation and knowledge articles Being a true proponent of customer advocacy Strengthening your AWS/Azure and Databricks platform expertise through learning and internal training programs Participating in weekend and weekday on call rotation</p>\n<p>Requirements include: Minimum 4 years experience designing, building, testing, and maintaining Python/Java/Scala based applications Expert level knowledge in python is desired Solid experience with SQL-based database is required Linux/Unix administration skills Hands-on experience with AWS, Azure or GCP Candidate must possess excellent English written and oral communication skills Experience with &quot;Distributed Big Data Computing&quot; environment Technical degree or the equivalent experience</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f2bc1be2-478","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7902994002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","Scala","SQL","Linux/Unix administration","AWS","Azure","GCP","Distributed Big Data Computing"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:52.913Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Rica"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Scala, SQL, Linux/Unix administration, AWS, Azure, GCP, Distributed Big Data Computing"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a33dd90a-a54"},"title":"Data Analyst, Corporate Operations","description":"<p>We&#39;re hiring a Data Analyst, Corporate Operations to support and scale Anduril&#39;s corporate planning and operating cadence through high-quality analytics, visualization, and quantitative analysis.</p>\n<p>In this role, you will translate planning and operational questions into metrics, dashboards, and analyses that inform decision-making. You will be responsible not only for producing rigorous analytical work, but also for ensuring the analytics products you build are reliable, well understood, and continuously improving based on how they are used.</p>\n<p>This role is ideal for an early-career analyst who is technically strong, thoughtful about system design, and motivated to improve how data is used across the organization.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build and maintain dashboards, scorecards, and analytical tools that track performance against plan and surface relevant risks, to be used in recurring planning and operating reviews</li>\n<li>Iterate on dashboards and analytical tools by incorporating feedback from users and evolving requirements as planning processes mature.</li>\n<li>Perform quantitative analyses including forecasting, trend analysis, scenario modeling, variance analysis, and root-cause investigations.</li>\n<li>Partner with data engineering and platform teams to define data requirements, validate data quality, and ensure analytical outputs are accurate and trusted.</li>\n<li>Document assumptions, definitions, and methodologies to ensure metrics are understood and applied consistently across the organization.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>1–2+ years of professional experience in an analytics-focused role (Data Analyst, Analytics Consultant, or similar).</li>\n<li>Degree in a quantitative or analytical field (Statistics, Engineering, Data Science, Math) or equivalent practical experience</li>\n<li>Strong foundation in advanced analytics; ability to write Python, R, SQL, or other analytic languages</li>\n<li>Experience building business-facing dashboards and visualizations using tools such as Palantir Foundry, Looker, Tableau, Power BI, or similar is a plus</li>\n<li>Demonstrated ability to work cross-functionally, aligning diverse stakeholders and driving outcomes</li>\n<li>Ability to thrive in a fast-paced ambiguous environment, balancing detail orientation with strategic judgment</li>\n<li>Must be authorized to work in the United States</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a33dd90a-a54","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5027763007","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":"$86,000-$114,000 USD","x-skills-required":["Python","R","SQL","Palantir Foundry","Looker","Tableau","Power BI","Data Analysis","Data Visualization","Quantitative Analysis"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:49.413Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, R, SQL, Palantir Foundry, Looker, Tableau, Power BI, Data Analysis, Data Visualization, Quantitative Analysis","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":86000,"maxValue":114000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e58b08f7-c31"},"title":"Senior Data Engineer","description":"<p>As a Senior Data Engineer on the Analytics Team, you will collaborate with stakeholders across the company to design, build and implement data pipelines and models that enable our next generation of technology to be deployed around the world. You will have a hand in helping shape the data platform vision at Anduril.</p>\n<p>We&#39;re looking for software and data engineers who are seeking high impact collaborative roles focused on driving operational execution. Ideally you are looking to learn what it takes to build the next generation of defence technology.</p>\n<p>Your responsibilities will include leading the design and roadmap for our data platform, partnering with operations, product, and engineering to advocate best practices and build supporting systems and infrastructure for the various data needs, owning the ingest and egress frameworks for data pipelines that stitch together various data sources in order to produce valuable data products that drive the business, and managing a large user base and providing true data self-service at scale.</p>\n<p>We use Palantir Foundry as our central hub for data-driven applications, visualizations and large-scale data analysis across the Anduril org. We also use SQLMesh for data transformations, Athena for querying data, Apache Iceberg as our table format, and Flyte for orchestration.</p>\n<p>Required qualifications include 5+ years of experience in a data engineering role building products, ideally in a fast-paced environment, good foundations in Python or another language, experience with Spark, PySpark, SQL and dbt, experience with Enterprise Data Systems like Palantir Foundry, and experience with or interest in learning how to develop data services and data products.</p>\n<p>The salary range for this role is $166,000-$220,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e58b08f7-c31","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/4587312007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$220,000 USD","x-skills-required":["Python","Spark","PySpark","SQL","dbt","Palantir Foundry","SQLMesh","Athena","Apache Iceberg","Flyte"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:44.003Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Spark, PySpark, SQL, dbt, Palantir Foundry, SQLMesh, Athena, Apache Iceberg, Flyte","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bdeae865-b3c"},"title":"Senior Software Engineer, Product","description":"<p>Join Brex, the intelligent finance platform that enables companies to spend smarter and move faster in over 200 markets. As a Senior Software Engineer, Product, you will help develop new products from concept to launch, building customer-facing products that drive direct business impact and shape our long-term technical vision with a high-quality bar.</p>\n<p>You will be responsible for owning functionality and scalability features by taking responsibility from inception to deployment, upholding our high engineering standards and bringing consistency to the codebases, infrastructure, and processes you will encounter.</p>\n<p>We&#39;re looking for folks with an interest in building products and tools, and who are comfortable in dealing with lots of moving pieces. You&#39;ll have the opportunity to learn and push the frontier of providing the best financial software experience to help companies grow.</p>\n<p>This role will be based in our Vancouver Office, which is a hybrid environment that combines the energy and connections of being in the office with the benefits and flexibility of working from home. We currently require a minimum of two coordinated days in the office per week, Wednesday and Thursday. Starting February 2, 2026, we will require 3 days per week in the office - Monday, Wednesday, and Thursday.</p>\n<p>As a perk, we also have up to four weeks per year of fully remote work!</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Own functionality and scalability features by taking responsibility from inception to deployment.</li>\n<li>Be responsible for full software development lifecycle - design, development, testing, operating in production.</li>\n<li>Uphold our high engineering standards and bring consistency to the codebases, infrastructure, and processes you will encounter.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>7+ years of professional experience in a software engineering role or equivalent experience.</li>\n<li>Experience building and designing scalable backend systems.</li>\n<li>Experience working with SQL or NoSQL databases.</li>\n<li>Experience working with backend programming languages (Java, Kotlin, Python).</li>\n<li>Familiarity with software engineering development cycles.</li>\n<li>Ability to hold yourself and the team to high standards.</li>\n<li>Strong communication and interpersonal skills.</li>\n<li>English proficiency/fluency, both written and speaking (note: interviews will be conducted in English).</li>\n</ul>\n<p>Bonus points:</p>\n<ul>\n<li>Experience collaborating with experts in product, design, and operations.</li>\n<li>Familiarity with functional programming languages.</li>\n<li>Experience driving initiatives at a broader level across an organization or company.</li>\n<li>Strong writing skills.</li>\n<li>Proactive approach.</li>\n</ul>\n<p>Compensation: The expected salary range for this role is $192,000 - $240,000 + equity. However, the starting base pay will depend on a number of factors including the candidate’s location, skills, experience, market demands, and internal pay parity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bdeae865-b3c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Brex","sameAs":"https://brex.com/","logo":"https://logos.yubhub.co/brex.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/brex/jobs/8398586002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$192,000 - $240,000 + equity","x-skills-required":["backend programming languages (Java, Kotlin, Python)","SQL or NoSQL databases","software engineering development cycles","high engineering standards","English proficiency/fluency"],"x-skills-preferred":["functional programming languages","collaborating with experts in product, design, and operations","driving initiatives at a broader level across an organization or company","strong writing skills","proactive approach"],"datePosted":"2026-04-18T15:58:42.816Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver, British Columbia, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"backend programming languages (Java, Kotlin, Python), SQL or NoSQL databases, software engineering development cycles, high engineering standards, English proficiency/fluency, functional programming languages, collaborating with experts in product, design, and operations, driving initiatives at a broader level across an organization or company, strong writing skills, proactive approach","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":192000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f86a39bf-9a5"},"title":"Solutions Architect - Digital Native Business, Strategic","description":"<p>As a Solutions Architect on the Digital Natives team, you will work with leading data engineering, data science, and ML teams to push the boundaries of what big data architectures are capable of.</p>\n<p>Reporting to the Field Engineering Manager, you will collaborate with strategic customers, product teams, and the broader customer-facing team to develop architectures and solutions using our platform and APIs.</p>\n<p>You will guide customers through the competitive landscape, best practices, and implementation; and develop technical champions along the way.</p>\n<p>We are looking for high technical aptitude individuals with a deep sense of ownership and a desire to help customers ship solutions at production scale.</p>\n<p>Ideal candidates are deeply curious, capable of operating with confidence in ambiguous situations, and are extremely adaptable.</p>\n<p>The impact you will have:</p>\n<ul>\n<li>Partner with the sales team and provide technical leadership to help customers understand how Databricks can help solve their business problems.</li>\n</ul>\n<ul>\n<li>Drive technical discovery and solution design, focusing on winning competitive deals and accelerating time-to-value in strategic accounts.</li>\n</ul>\n<ul>\n<li>Continuously research &amp; learn new technologies and their implementations on Databricks.</li>\n</ul>\n<ul>\n<li>Consult on Big Data architectures, implement proof of concepts for strategic projects, spanning data engineering, data science, and machine learning, and SQL analysis workflows.</li>\n</ul>\n<ul>\n<li>As well as validating integrations with cloud services, home-grown tools, and other 3rd party applications.</li>\n</ul>\n<ul>\n<li>Collaborate with your fellow Solutions Architects, using your skills to support each other and our customers.</li>\n</ul>\n<ul>\n<li>Become an expert in, promote, and recruit contributors for Databricks-inspired open-source projects (Spark, Delta Lake, and MLflow) across the developer community.</li>\n</ul>\n<ul>\n<li>Work closely with account executives to create and execute account penetration strategies, focusing on winning technical decision-makers and building new customer champions.</li>\n</ul>\n<ul>\n<li>Build trusted advisor relationships with senior and executive stakeholders by articulating the business value of Databricks in clear, outcomes-driven terms.</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>5+ years in a data engineering, data science, technical architecture, or similar pre-sales/consulting role.</li>\n</ul>\n<ul>\n<li>Experience building distributed data systems.</li>\n</ul>\n<ul>\n<li>Comfortable programming in, and debugging, Python and SQL.</li>\n</ul>\n<ul>\n<li>Have built solutions with public cloud providers such as AWS, Azure, or GCP.</li>\n</ul>\n<ul>\n<li>Expertise in one of the following:</li>\n</ul>\n<ul>\n<li>Data Engineering technologies (Ex: Spark, Hadoop, Kafka)</li>\n</ul>\n<ul>\n<li>Data Science and Machine Learning technologies (Ex: pandas, scikit-learn, pytorch, Tensorflow)</li>\n</ul>\n<ul>\n<li>Strong executive presence with the ability to influence C/VP-level stakeholders and align technical solutions to strategic business priorities.</li>\n</ul>\n<ul>\n<li>Available to travel to customers in your region.</li>\n</ul>\n<ul>\n<li>[Desired] Degree in a quantitative discipline (Computer Science, Applied Mathematics, Operations Research).</li>\n</ul>\n<ul>\n<li>Nice to have: Databricks Certification.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f86a39bf-9a5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8434467002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,000-$247,500 USD","x-skills-required":["Data Engineering technologies","Data Science and Machine Learning technologies","Python","SQL","Cloud providers (AWS, Azure, GCP)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:42.812Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - California; Remote - Colorado; Remote - Oregon; Remote - Washington"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Data Engineering technologies, Data Science and Machine Learning technologies, Python, SQL, Cloud providers (AWS, Azure, GCP)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":247500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7eae126b-eb2"},"title":"Senior Accountant","description":"<p>We&#39;re seeking a highly motivated Senior Accountant to join our growing accounting team. As a Senior Accountant, you will assist in preparing day-to-day corporate accounting operations, supporting the month-end close process, and helping to implement systems and processes that will support Scale as we continue to grow.</p>\n<p>The ideal candidate thrives in a high-growth start-up, is detail-oriented, and has excellent interpersonal and communication skills. Additionally, the candidate has demonstrated the ability to build scalable cross-functional relationships through systems and process implementation.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Prepare journal entries and day-to-day corporate accounting activities, support the month-end close process, and provide timely and accurate month-end close financials that are U.S. GAAP compliant</li>\n<li>Build or enhance balance sheet account reconciliation workpaper including reviewing and performing some clean-up of historical reconciliations and related balances</li>\n<li>Collaborate within Accounting and Finance teams on metrics, flux analysis, forecast, and projections and support preparation of the monthly reporting package</li>\n<li>Prepare documents supporting internal and external audits and ensure the successful completion of those audits</li>\n<li>Support the implementation of new systems, tools, and processes to streamline close and build scalable solutions to support the growth of the Company</li>\n<li>Identify and drive process improvements to gain efficiencies and reduce close timeline</li>\n<li>Develop, maintain and improve internal controls which relate to assigned areas</li>\n</ul>\n<p>Ideally You Have:</p>\n<ul>\n<li>Bachelor’s degree in Accounting; CPA or in the process of working towards one</li>\n<li>3+ years of relevant accounting experience; Combination of public accounting (Big 4 strongly preferred) and industry experience (start-up through to becoming a public Company experience a plus)</li>\n<li>Strong knowledge of U.S. GAAP and SOX</li>\n<li>Experience with general ledger functions and the month-end/year-end close process is strongly preferred</li>\n<li>Experience with financial audit preparation is strongly preferred</li>\n<li>Proven proficiency of Microsoft Office Suite (Word, PowerPoint, etc); must be highly proficient in Excel</li>\n<li>Excellent writing, oral communication, analytical, organizational, and interpersonal skills with a high attention to detail</li>\n<li>Excellent problem-solving skills; project management experience a plus, ability to work in a fast-paced environment</li>\n</ul>\n<p>Nice to Haves:</p>\n<ul>\n<li>Direct accounting experience for any/all of the following areas: accruals, investment, internally used software, ASC 842 leases, cloud computing arrangement, stock-based compensation and/or consolidation and Fx.</li>\n<li>Proficiency in SQL, NetSuite, FloQast, Coupa, and Jira considered a plus</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity-based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p>About Us:</p>\n<p>At Scale, our mission is to develop reliable AI systems for the world&#39;s most important decisions. Our products provide the high-quality data and full-stack technologies that power the world&#39;s leading models, and help enterprises and governments build, deploy, and oversee AI applications that deliver real impact.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7eae126b-eb2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4663533005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$92,000-$115,000 USD","x-skills-required":["U.S. GAAP","SOX","Microsoft Office Suite","Excel","General Ledger Functions","Month-End/Year-End Close Process"],"x-skills-preferred":["SQL","NetSuite","FloQast","Coupa","Jira"],"datePosted":"2026-04-18T15:58:41.071Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Technology","skills":"U.S. GAAP, SOX, Microsoft Office Suite, Excel, General Ledger Functions, Month-End/Year-End Close Process, SQL, NetSuite, FloQast, Coupa, Jira","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":92000,"maxValue":115000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6683dad6-0d9"},"title":"Senior Data Scientist","description":"<p>Join us to build the decision engine for better mental health outcomes.</p>\n<p>As a Senior Data Scientist, you will sit in the heart of a cross-functional product team and help turn messy, real-world signals into clear decisions. You will make sure we are capturing the right data, designing experiments that tell us what is actually driving outcomes, and translating findings into recommendations that teams can act on quickly.</p>\n<p>When the insight is stable and valuable, you will help operationalize it through predictive models that improve provider and patient experiences.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Being the analytics partner inside the pod, working closely with Product, Engineering, Design, Ops, and Clinical stakeholders to define questions, metrics, guardrails, and decision rules.</li>\n<li>Running rigorous experiments, designing and analyzing A/B tests and quasi-experiments with clear hypotheses, power considerations, and pre-defined success criteria.</li>\n<li>Connecting behavior to strategy, using funnel, cohort, segmentation, and lifecycle analysis to understand how people and providers experience Headway, and where product changes will have the biggest impact.</li>\n<li>Using causal inference when experiments are not possible, applying approaches like diff-in-diff, matching, and regression-based designs with principled uncertainty quantification.</li>\n<li>Building models when they should exist, developing predictive models that operationalize vetted insights (feature development, validation, backtesting, calibration), with clear launch criteria and monitoring plans.</li>\n<li>Creating decision-ready work, producing analysis and narratives that are crisp, honest about uncertainty, and drive action.</li>\n</ul>\n<p>To be successful in this role, you will need:</p>\n<ul>\n<li>6+ years using data to drive product or business decisions in product, growth, engineering, or operations environments.</li>\n<li>Strong SQL and strong proficiency in Python or R for analysis and modeling.</li>\n<li>Demonstrated depth in experimentation and causal inference under real-world constraints.</li>\n<li>Practical modeling skill: feature engineering, model comparison, cross-validation or backtesting, calibration, and post-launch monitoring.</li>\n<li>Strong product sense and opinions, including a track record of connecting analytics recommendations to measurable outcomes.</li>\n<li>Clear communication: you can explain complex work to non-technical audiences without losing the truth.</li>\n<li>A self-starter mindset: you prioritize well, follow through, and do not need heavy oversight.</li>\n<li>Motivation for our mission: improving access and affordability in mental healthcare.</li>\n</ul>\n<p>The expected base pay range for this position is $180,000 - $225,000, based on a variety of factors including qualifications, experience, and geographic location. In addition to base salary, this role may be eligible for an equity grant, depending on the position and level.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6683dad6-0d9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Headway","sameAs":"https://headway.co/","logo":"https://logos.yubhub.co/headway.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/headway/jobs/5677823004","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,000 - $225,000","x-skills-required":["SQL","Python","R","Experimentation","Causal Inference","Predictive Modeling","Data Analysis","Communication"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:38.344Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States; San Francisco, California, United States; Seattle, Washington, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"SQL, Python, R, Experimentation, Causal Inference, Predictive Modeling, Data Analysis, Communication","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":225000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a1f35a9c-2e5"},"title":"Staff Data Scientist","description":"<p>We are looking for a Staff Product Data Scientist to join our Data &amp; Insights team. This role will lead the strategy, development, and operationalization of advanced analytics and machine learning solutions that power data intelligence for Okta’s Product Management teams, uncovering actionable insights on customer behavior, product engagement, and opportunities that will drive our product growth strategy.</p>\n<p>You’ll define the long-term data science roadmap while partnering closely with Okta’s diverse Product Management teams and executive leadership.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Collaborate closely with Product Managers and leadership to provide insights that will help drive product strategy and roadmaps, ensuring decisions are grounded in a data-driven approach.</li>\n<li>Build and deploy statistical and machine learning models (e.g., predicting churn, lifetime value, feature adoption) to forecast user behavior and product growth opportunities in order to help influence roadmap decisions.</li>\n<li>Conduct thorough exploratory analysis on extensive, complex datasets to uncover key drivers of user adoption and engagement, identifying unseen opportunities for significant product improvement.</li>\n<li>Work with data engineering, analytics engineering, and data analysts to shape and enhance the foundational data infrastructure necessary for scalable ML and advanced analytics initiatives.</li>\n<li>Develop and construct usable data sets by integrating and manipulating information from various disparate data sources as needed.</li>\n<li>Translate intricate data insights into clear, compelling narratives for executives, product managers, and engineers, effectively influencing crucial business and product decisions.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>7+ years of experience in data science or ML, including 3–5 years in a senior, staff, or technical leadership role.</li>\n<li>Deep understanding of product analytics and data science frameworks, with a proven history of designing and analyzing product data to solve complex, ambiguous business problems and deliver measurable results.</li>\n<li>Demonstrated ability to apply cutting-edge AI tools to accelerate the discovery of deep, actionable insights from complex product data.</li>\n<li>Demonstrated ability to translate data insights into product impact and formulate strategic, data-driven recommendations.</li>\n<li>Deep expertise in machine learning algorithms (supervised, unsupervised, NLP, forecasting, optimization) and statistical modeling.</li>\n<li>Strong proficiency in Python, SQL, and leading ML libraries.</li>\n<li>Excellent communication skills and a proven ability to influence Product and executive partners.</li>\n</ul>\n<p><strong>Nice to Have</strong></p>\n<ul>\n<li>Experience in high-growth SaaS, cybersecurity, identity, or enterprise software environments.</li>\n<li>Prior ownership of ML platform/tooling decisions and evaluations.</li>\n<li>Experience enabling self-service analytics or citizen data science capabilities.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a1f35a9c-2e5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7731595","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$184,000-$253,000 USD","x-skills-required":["data science","machine learning","Python","SQL","statistical modeling","data engineering","analytics engineering","data analysts"],"x-skills-preferred":["high-growth SaaS","cybersecurity","identity","enterprise software"],"datePosted":"2026-04-18T15:58:37.181Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bellevue, Washington; San Francisco, California"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data science, machine learning, Python, SQL, statistical modeling, data engineering, analytics engineering, data analysts, high-growth SaaS, cybersecurity, identity, enterprise software","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":184000,"maxValue":253000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b02d4cf4-599"},"title":"Director, Business Systems","description":"<p>About the Role</p>\n<p>The Director of Business Systems will be responsible for building real depth across Finance, GTM, Operations, and People systems, connecting these capabilities into a powerful, AI-integrated backbone.</p>\n<p>You are a builder-first leader who will blend enterprise architecture with cutting-edge AI automation. Your goal is to eliminate fragmentation and enable every function to move faster by deploying internal AI agents that reduce manual work and surface actionable insights.</p>\n<p>Key Responsibilities</p>\n<ul>\n<li>Enterprise Application Strategy: Own the full lifecycle of our application ecosystem, including ERP (Finance), CRM (GTM), HRIS (People), and specialized AI tools. Ensure these systems converge into a secure, scalable backbone.</li>\n</ul>\n<ul>\n<li>AI Agent Deployment: Identify opportunities to replace manual workflows with AI/LLM-powered agents. Build and manage &#39;internal agents&#39; that automate forecasting, revenue capture, and employee self-service.</li>\n</ul>\n<ul>\n<li>Cross-Functional Partnership: Serve as the primary technology partner to the leadership team. Align system roadmaps with departmental goals to ensure seamless data flow from production to the back office.</li>\n</ul>\n<ul>\n<li>Operational Excellence: Mature our core systems by establishing disciplined change management, clear data governance, and measurable SLAs. Stabilize environments to support audit and regulatory requirements.</li>\n</ul>\n<ul>\n<li>Team Leadership: Lead a high-impact team that fosters a culture of curiosity, speed, and user-centricity.</li>\n</ul>\n<p>Ideally, you have:</p>\n<ul>\n<li>10+ years of experience leading business systems or enterprise applications in a fast-paced, high-growth environment.</li>\n</ul>\n<ul>\n<li>Technical Depth: Strong functional understanding of Salesforce (or equivalent CRM), NetSuite (or equivalent ERP), and Workday (or equivalent HRIS).</li>\n</ul>\n<ul>\n<li>AI/Automation Mindset: Proven track record of designing and deploying AI/LLM-powered workflows or agentic systems to improve business efficiency.</li>\n</ul>\n<ul>\n<li>Architecture Skills: Ability to design scalable data flows and APIs that link disparate SaaS tools into a cohesive ecosystem.</li>\n</ul>\n<ul>\n<li>Stakeholder Mastery: Experience influencing C-level executives and translating complex technical needs into business outcomes.</li>\n</ul>\n<ul>\n<li>Skilled in Python or SQL to personally audit or prototype automation logic.</li>\n</ul>\n<ul>\n<li>Extensive experience and In-depth knowledge of the functionality of the ERP modules with emphasis on Sales – Order to Cash &amp; Sales Audit, Purchasing – Direct &amp; Indirect and Finance - General Ledger, AP &amp; AR, Procure to Pay, Sales Audit, Fixed Assets, International Consolidations &amp; Reporting.</li>\n</ul>\n<ul>\n<li>Experience with GTM systems such as, HubSpot, Outreach, Clari, CPQ (e.g., Salesforce CPQ), and similar tools.</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p>You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p>The base salary range for this full-time position in the location of San Francisco is: $235,200-$294,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b02d4cf4-599","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4657910005","x-work-arrangement":"onsite","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":"$235,200-$294,000 USD","x-skills-required":["Salesforce","NetSuite","Workday","Python","SQL","AI/LLM-powered workflows","Scalable data flows","APIs","Disparate SaaS tools","C-level executives","Business outcomes","ERP modules","GTM systems","HubSpot","Outreach","Clari","CPQ"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:33.977Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Salesforce, NetSuite, Workday, Python, SQL, AI/LLM-powered workflows, Scalable data flows, APIs, Disparate SaaS tools, C-level executives, Business outcomes, ERP modules, GTM systems, HubSpot, Outreach, Clari, CPQ","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":235200,"maxValue":294000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3ba73370-831"},"title":"Internal Audit IT Manager","description":"<p>Ready to be pushed beyond what you think you’re capable of?</p>\n<p>At Coinbase, our mission is to increase economic freedom in the world.</p>\n<p>We’re seeking a very specific candidate who is passionate about our mission and who believes in the power of crypto and blockchain technology to update the financial system.</p>\n<p>As an Internal Audit IT Manager, you will own end-to-end delivery of complex IT and security audits across our cloud infrastructure, security operations, and crypto-native systems.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Owning end-to-end delivery of IT and security audits, from risk assessment and scoping through planning, fieldwork, testing, reporting, and issue validation,covering cloud infrastructure (AWS, GCP), security operations, identity and access management, data protection, IT asset management, vendor/third-party risk, and key in-scope products and services including blockchain infrastructure, centralized and self-hosted wallets, and cold storage.</li>\n</ul>\n<ul>\n<li>Driving AI-enabled audit execution, designing and implementing data analytics, automation, and Generative AI solutions to modernize how we audit (e.g., continuous monitoring, anomaly detection, automated evidence retrieval, AI-assisted workpaper drafting),while maintaining rigorous human-in-the-loop validation to ensure accuracy and audit-quality conclusions.</li>\n</ul>\n<ul>\n<li>Executing audits aligned with the multi-year IT and security audit roadmap, coordinating coverage with co-sourced partners and cross-functional risk initiatives while ensuring alignment with Coinbase&#39;s enterprise risk profile, technology strategy, and regulatory expectations across regions (US, EMEA, APAC).</li>\n</ul>\n<ul>\n<li>Driving high-quality, risk-based findings and executive-level reporting, distilling key themes, emerging risks, and root causes into clear, concise materials for senior management and the Chief Audit Executive,ensuring findings are appropriately documented and supported by evidence.</li>\n</ul>\n<ul>\n<li>Partnering with technology and security leadership across Engineering, Security, Infrastructure, Product, and Operations to build trusted relationships, challenge control design, and advise on pragmatic, risk-based, scalable remediation while maintaining third-line independence.</li>\n</ul>\n<ul>\n<li>Driving disciplined issue management, ensuring timely, risk-based remediation by management, high-quality root cause analysis, and validation of remediation activities,escalating delays or thematic concerns to senior leadership as needed.</li>\n</ul>\n<ul>\n<li>Evaluating and developing talent, assessing candidates and helping build a high-performing, technically credible audit team.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>7+ years of experience in IT/security internal audit, technology risk, or first-line security/engineering roles with significant controls exposure.</li>\n</ul>\n<ul>\n<li>Experience working in a fast-paced, cloud-native, or engineering-driven environment where technology and security practices evolve rapidly.</li>\n</ul>\n<ul>\n<li>Hands-on audit experience with cloud platforms (AWS, GCP), including IAM policies, security configurations, logging/monitoring, and CI/CD pipelines.</li>\n</ul>\n<ul>\n<li>AI-forward mindset with demonstrated experience applying Python, SQL, or AI tools to audit or security work, building workflows rather than just prompting.</li>\n</ul>\n<ul>\n<li>Relevant professional certifications (e.g., CISA, CISSP, CIA, CISM) required; CPA or CFE a plus.</li>\n</ul>\n<ul>\n<li>Working knowledge of key frameworks such as NIST CSF, COBIT, SOC 2, and ITIL.</li>\n</ul>\n<ul>\n<li>High EQ and collaborative style.</li>\n</ul>\n<ul>\n<li>Proven ability to translate complex technical findings into clear, executive-ready narratives for both technical and non-technical audiences.</li>\n</ul>\n<ul>\n<li>Ability to manage multiple audits and initiatives across time zones (EMEA, APAC) with minimal oversight.</li>\n</ul>\n<ul>\n<li>Demonstrated leadership and team-development experience, including mentoring, coaching, and managing direct reports.</li>\n</ul>\n<ul>\n<li>Demonstrates the ability to responsibly use generative AI tools and copilots (e.g., LibreChat, Gemini, Glean) in daily workflows, continuously learn as tools evolve, and apply human-in-the-loop practices to deliver business-ready outputs and drive measurable improvements in efficiency, cost, and quality.</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience auditing or building blockchain infrastructure, crypto custody, or wallet systems (hot/cold storage).</li>\n</ul>\n<ul>\n<li>Background in a high-growth or rapidly scaling environment with complex, evolving technology stacks.</li>\n</ul>\n<ul>\n<li>Experience with GRC platforms (Workiva, Archer, AuditBoard) or building custom audit automation tooling.</li>\n</ul>\n<ul>\n<li>Familiarity with DORA, MiCA, or crypto-specific regulatory frameworks.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3ba73370-831","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Coinbase","sameAs":"https://www.coinbase.com/","logo":"https://logos.yubhub.co/coinbase.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coinbase/jobs/7755116","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,345-$195,700 USD","x-skills-required":["IT security","Cloud infrastructure","Security operations","Identity and access management","Data protection","IT asset management","Vendor/third-party risk","Blockchain infrastructure","Centralized and self-hosted wallets","Cold storage","AI-enabled audit execution","Data analytics","Automation","Generative AI","Continuous monitoring","Anomaly detection","Automated evidence retrieval","AI-assisted workpaper drafting","Cloud platforms","IAM policies","Security configurations","Logging/monitoring","CI/CD pipelines","Python","SQL","AI tools","NIST CSF","COBIT","SOC 2","ITIL","CISA","CISSP","CIA","CISM","CPA","CFE"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:31.708Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"IT security, Cloud infrastructure, Security operations, Identity and access management, Data protection, IT asset management, Vendor/third-party risk, Blockchain infrastructure, Centralized and self-hosted wallets, Cold storage, AI-enabled audit execution, Data analytics, Automation, Generative AI, Continuous monitoring, Anomaly detection, Automated evidence retrieval, AI-assisted workpaper drafting, Cloud platforms, IAM policies, Security configurations, Logging/monitoring, CI/CD pipelines, Python, SQL, AI tools, NIST CSF, COBIT, SOC 2, ITIL, CISA, CISSP, CIA, CISM, CPA, CFE","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166345,"maxValue":195700,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4daeb1d2-f04"},"title":"Senior Software Engineer - Fullstack","description":"<p>We are seeking a senior software engineer to join our team in Vancouver. As a fullstack software engineer, you will work with your team and product management to make insights from data simple. You&#39;ll set the foundation for how we build robust, scalable, and delightful products.</p>\n<p>Our customers increasingly use Databricks to analyze petabyte-scale logs in real time. This creates new challenges across the entire data processing pipeline, including ingestion, indexing, processing, and the user experience itself. Our customers are also using Databricks to launch AI/BI, which is redefining Business Intelligence for the AI age. We have several open roles across the teams below:</p>\n<ul>\n<li>Log Analytics: Our customers increasingly use Databricks to analyze petabyte-scale logs in real time.</li>\n<li>AI/BI: AI/BI is redefining Business Intelligence for the AI age.</li>\n<li>Unity Catalog Business Semantics: Context is everything for AI. For enterprise data, that context needs to be governed and managed, which is what Unity Catalog Business Semantics offers.</li>\n<li>Databricks Apps: Databricks Apps is one of the fastest growing products at Databricks, used by more than 2,500 customers who have created more than 20,000 apps.</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>5+ years of experience with HTML, CSS, and JavaScript.</li>\n<li>Passion for user experience and design and a deep understanding of front-end architecture.</li>\n<li>Comfortable working towards a multi-year vision with incremental deliverables.</li>\n<li>Motivated by delivering customer value.</li>\n<li>Experience with modern JavaScript frameworks (e.g., React, Angular, or VueJs/Ember).</li>\n<li>5+ years of experience with server-side web technologies (eg: Node.js, Java, Python, Scala, C#, C++,Go).</li>\n<li>Good knowledge of SQL.</li>\n<li>Experience with cloud technologies, e.g. AWS, Azure, GCP, Docker, or Kubernetes.</li>\n<li>Experience developing large-scale distributed systems.</li>\n</ul>\n<p>Pay Range Transparency Databricks is committed to fair and equitable compensation practices. The pay range for this role is listed below and represents the expected salary range for non-commissionable roles or on-target earnings for commissionable roles. Actual compensation packages are based on several factors that are unique to each candidate, including but not limited to job-related skills, depth of experience, relevant certifications and training, and specific work location. Based on the factors above, Databricks anticipates utilizing the full width of the range. The total compensation package for this position may also include eligibility for annual performance bonus, equity, and the benefits listed above. Canada Pay Range $146,200-$201,100 CAD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4daeb1d2-f04","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8099342002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$146,200-$201,100 CAD","x-skills-required":["HTML","CSS","JavaScript","Node.js","Java","Python","Scala","C#","C++","Go","SQL","AWS","Azure","GCP","Docker","Kubernetes"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:30.534Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"HTML, CSS, JavaScript, Node.js, Java, Python, Scala, C#, C++, Go, SQL, AWS, Azure, GCP, Docker, Kubernetes","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":146200,"maxValue":201100,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bd128020-804"},"title":"Consumer Revenue Strategy & Operations Manager","description":"<p>We are seeking a Consumer Revenue Strategy &amp; Operations Manager to own strategy and operations for Discord&#39;s Consumer Revenue business, with a primary focus on Nitro partnerships. This is a high-impact role that touches deal strategy, financial modeling, pricing, analytics, and cross-functional operations.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Serve as the strategic connective tissue between Business Development and Product on Nitro partnerships, ensuring alignment on deal priorities, product capabilities, and go-to-market strategy</li>\n<li>Provide analytical support to Consumer Revenue leadership by building reporting, surfacing insights on partnership performance, and delivering data-driven recommendations that inform business decisions</li>\n<li>Partner with Finance on financial modeling for Nitro partnership deals, building and maintaining models that inform deal structure, revenue projections, and partnership economics</li>\n<li>Stand up and run the Consumer Revenue deal desk, defining red lines on term sheets, establishing approval frameworks, and creating clear guardrails on what we will and won&#39;t allow in partnership agreements</li>\n<li>Lead pricing strategy and analysis for Nitro, managing the semi-annual pricing cycle and building the analytical foundation to support pricing decisions</li>\n<li>Develop and maintain dashboards and ad hoc analyses that give Business Development and Product teams visibility into partner revenue trends, deal pipeline health, and key performance metrics</li>\n<li>Manage operational workflows across Accounting, Legal, and other cross-functional partners to ensure deals move from signed term sheet to revenue recognition smoothly</li>\n<li>Document and codify deal policies, pricing frameworks, and operational processes as the function matures</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>4 to 6 years of experience in strategy and operations, deal operations, revenue operations, or strategic finance</li>\n<li>Strong financial modeling skills with the ability to build deal models, scenario analysis, and partnership economics from scratch</li>\n<li>Proven ability to turn ambiguous business questions into structured analyses and clear recommendations</li>\n<li>Experience standing up or managing a deal desk function, including defining deal terms, approval workflows, and pricing guardrails</li>\n<li>Ability to operate across multiple stakeholders (Business Development, Product, Marketing, Engineering, Legal, Finance, and Accounting) and drive alignment without direct authority</li>\n<li>Sharp analytical skills with proficiency with SQL, python, BI Tools, and Excel/Sheets.</li>\n<li>Proficient with AI tools such as Claude, Claude Code, Cursor, etc</li>\n<li>Strong written and verbal communication.</li>\n</ul>\n<p>Nice to Have:</p>\n<ul>\n<li>Experience in subscription or consumer revenue businesses, or partnership-heavy models</li>\n<li>Background in gaming, social platforms, or digital media</li>\n<li>Familiarity with revenue recognition principles and working with Accounting teams on deal operationalization</li>\n<li>Experience with BI tools such as Tableau, Looker, or similar, and building self-serve reporting for business teams</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bd128020-804","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Discord","sameAs":"https://discord.com/","logo":"https://logos.yubhub.co/discord.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/discord/jobs/8507387002","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$192,000 to $216,000 + equity + benefits","x-skills-required":["financial modeling","deal operations","revenue operations","strategic finance","SQL","python","BI Tools","Excel/Sheets","AI tools","Claude","Claude Code","Cursor"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:29.536Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco Bay Area"}},"employmentType":"FULL_TIME","occupationalCategory":"Operations","industry":"Technology","skills":"financial modeling, deal operations, revenue operations, strategic finance, SQL, python, BI Tools, Excel/Sheets, AI tools, Claude, Claude Code, Cursor","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":192000,"maxValue":216000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c1d056a0-ebf"},"title":"Staff Software Engineer, Reporting Platform","description":"<p>About Gusto</p>\n<p>At Gusto, we&#39;re on a mission to grow the small business economy. We handle the hard stuff , payroll, health insurance, 401(k)s, and HR , so owners can focus on their craft and their customers.</p>\n<p>The Reporting Platform team at Gusto empowers business owners to make better decisions with data and insights through reports and visualizations that span our product lines. As a member of our Reporting Platform team, you will create and maintain reports and enable engineering teams to deposit and consume data from our reporting platform.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Collaboratively design and implement reports and visualizations across Gusto’s product suite in our Ruby on Rails/React-based stack.</li>\n<li>Migrate reports from our legacy infrastructure (Rails/MySQL) to our new reporting platform (Rails/Cube/Clickhouse).</li>\n<li>Frequently demonstrate your work to your team and the broader engineering organization.</li>\n<li>Improve the quality of our offerings by participating in support rotations and maintaining a prioritized backlog of technical debt and SRE improvements.</li>\n<li>Lead and mentor fellow engineers in tackling complex technical challenges at scale.</li>\n<li>Prototype, iterate, and launch new features quickly and efficiently.</li>\n<li>Foster a collaborative environment that encourages creativity and innovation, building products our customers love.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>7+ years of professional software development experience.</li>\n<li>Highly proficient in HTML, CSS, JavaScript, React, TypeScript.</li>\n<li>Very strong understanding of SaaS fundamentals.</li>\n<li>Excellent communicator.</li>\n<li>Willingness to learn new domains and quickly develop expertise.</li>\n</ul>\n<p>Total Rewards</p>\n<p>Our cash compensation amount for this role is targeted at $200,000/yr to $247,000/yr for New York City. Stock equity is additional.</p>\n<p>Work Environment</p>\n<p>Gusto has physical office spaces in Denver, San Francisco, and New York City. Employees who are based in those locations will be expected to work from the office on designated days approximately 2-3 days per week (or more depending on role).</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c1d056a0-ebf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Gusto","sameAs":"https://www.gusto.com/","logo":"https://logos.yubhub.co/gusto.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gusto/jobs/7654894","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$200,000/yr to $247,000/yr","x-skills-required":["HTML","CSS","JavaScript","React","TypeScript","Ruby on Rails","MySQL","Cube","Clickhouse"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:29.419Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"HTML, CSS, JavaScript, React, TypeScript, Ruby on Rails, MySQL, Cube, Clickhouse","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":200000,"maxValue":247000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7cbdd2a7-0be"},"title":"Strategic Project Lead, Robotics","description":"<p>Scale&#39;s Robotics business unit is experiencing historic levels of growth.</p>\n<p>As a Strategic Projects Lead (SPL) for the Data Collection team, you will lead initiatives that drive $XXM+ in new revenue for the business.</p>\n<p>This is a demanding role, and as an SPL, you should be prepared to wear many hats across all dimensions of Operations. The ideal SPL has a strong entrepreneurial mindset, is comfortable getting into the weeds, and is excited about intense, impactful work that leads to accelerated career progression.</p>\n<p>Key Responsibilities:</p>\n<p>Serve as the full owner of our most visible and high-impact customer and internal collection projects, making decisions that directly impact data quality, operational efficiency, revenue, and margins. Understand customer requirements and design collection specifications best suited to improving model performance based on customer needs. Build out project timelines and pipeline infrastructure to ensure quality and efficiency. Train, coach, and manage dynamic and global teams of operators and collectors. Build analytics and reports to make data-driven decisions. Functionally lead a team of Operations Program Managers and Associates that support the execution of the project, define and assign tasks, monitor progress. Partner with diverse stakeholders (Engineering + Product + Finance + Go-to-Market) to work on problems that will drive advancements for the largest Robotics models (VLAs/VLMs) in the world. Give regular progress updates to Scale&#39;s executive team.</p>\n<p>Ideal Qualifications:</p>\n<p>Strong technical background, combined with a business-oriented mindset (a degree in engineering or computer science is ideal + prior exposure to driving business outcomes). 3+ years of experience leading a project team, developing product or operational processes, or launching new operations. Excellent problem-solving capabilities (experience working on operational challenges is a plus). Ability to come up with creative solutions to complex, ambiguous operational problems. Entrepreneurial experience and mindset - you are excited about building things from scratch and are able to identify issues and execute quickly. Nice to haves: Hands-on experience in Robotics, AI, and/or Computer Vision. Experience working with global teams on different timezones. SQL proficiency.</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity-based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend. Please reference the job posting&#39;s subtitle for where this position will be located. For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is: $152,000-$190,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7cbdd2a7-0be","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4658316005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$152,000-$190,000 USD","x-skills-required":["strong technical background","business-oriented mindset","project management","problem-solving","entrepreneurial experience"],"x-skills-preferred":["Robotics","AI","Computer Vision","SQL"],"datePosted":"2026-04-18T15:58:22.818Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"strong technical background, business-oriented mindset, project management, problem-solving, entrepreneurial experience, Robotics, AI, Computer Vision, SQL","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":152000,"maxValue":190000,"unitText":"YEAR"}}}]}