Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Qualifications
To excel in this role, you should have a strong background in software engineering and a passion for AI technologies. Key qualifications include:Bachelor's degree in Computer Science or a related fieldExperience with machine learning frameworks and tools (e.g., TensorFlow, PyTorch)Proficiency in programming languages such as Python, Java, or C++Strong analytical and problem-solving skillsExcellent communication and teamwork abilities
About the job
Join soum as an AI / GenAI Solutions Engineer and contribute to cutting-edge projects that leverage artificial intelligence and generative AI technologies. In this pivotal role, you will be responsible for designing, developing, and implementing innovative AI solutions that enhance our product offerings and drive customer satisfaction.
As part of our dynamic team, you will collaborate with cross-functional departments to gather requirements, develop prototypes, and ensure seamless integration of AI features into existing systems. You will also have the opportunity to stay abreast of the latest industry trends and advancements in AI technology.
About soum
soum is a forward-thinking company dedicated to transforming industries through innovative AI solutions. With a focus on collaboration and continuous learning, we foster an inclusive environment where creativity and technical expertise thrive. Join us and be part of a team that is shaping the future of technology.
Join soum as an AI / GenAI Solutions Engineer and contribute to cutting-edge projects that leverage artificial intelligence and generative AI technologies. In this pivotal role, you will be responsible for designing, developing, and implementing innovative AI solutions that enhance our product offerings and drive customer satisfaction.As part of our dynamic …
ABOUT PLACER.AI: Placer.ai is revolutionizing the way organizations comprehend the physical world. Our state-of-the-art location analytics platform offers unparalleled insights into locations, markets, and consumer behavior. By empowering a diverse range of customers—from Fortune 500 corporations to local governments and nonprofits—Placer enables data-driven decision-making. What distinguishes us? We have developed the most advanced location intelligence platform available while upholding an unwavering commitment to privacy. This demonstrates that powerful analytics can coexist with responsible data practices. Our rapid growth reflects the market's enthusiasm: in just six years since our inception, we achieved $100 million in annual recurring revenue, reached unicorn status with a valuation exceeding $1 billion in 2022, and continue to expand rapidly as one of North America's fastest-growing tech companies. We are creating a market opportunity worth over $100 billion and are only just getting started. Recognized as one of Forbes America's Best Startup Employers and a Deloitte Technology Fast 500 company, we are cultivating a culture where innovation flourishes, collaboration is paramount, and every team member plays a vital role in transforming how the world perceives location. SUMMARY: We are on the lookout for a seasoned Principal Engineer with extensive expertise in Generative AI to spearhead our GenAI product solutions and development initiatives. In this pivotal role, you will drive the technical vision and architecture of our AI-driven location intelligence platform, developing production-ready solutions that reshape how businesses harness and interpret location data. You will manage the full-stack development of our conversational AI system, encompassing LLM integration, agentic workflows, and real-time streaming interfaces, while laying the technical groundwork for our upcoming business intelligence tools. If you are a technical leader with over 10 years of software development experience, a proven track record in production GenAI settings, and a fervor for creating scalable AI platforms that yield tangible business benefits, we want to connect with you.
ABOUT PLACER.AI: At Placer.ai, we are revolutionizing the way organizations perceive and analyze the physical world. Our cutting-edge location analytics platform offers unmatched insights into locations, market trends, and consumer behaviors. Placer enables thousands of clients—from Fortune 500 companies to local governments and nonprofits—to make informed, data-driven decisions. What differentiates us? We have developed the most advanced location intelligence platform available while upholding an unwavering commitment to privacy, demonstrating that impactful analytics and ethical data practices can indeed coexist. Our rapid growth is a testament to the market's demand: we reached $100 million in annual recurring revenue within just six years of our inception, achieved unicorn status with a valuation exceeding $1 billion in 2022, and are consistently expanding as one of North America's fastest-growing tech firms. We are on the brink of creating a market opportunity exceeding $100 billion, and we are just getting started. Proudly recognized as one of Forbes' Best Startup Employers and a Deloitte Technology Fast 500 company, we are fostering a culture where innovation flourishes, collaboration thrives, and every member plays a pivotal role in redefining the understanding of location. SUMMARY: We are in search of a highly skilled Principal Engineer with extensive expertise in Generative AI to spearhead our GenAI product solutions and development initiatives. In this pivotal role, you will be responsible for shaping the technical vision and architecture of our AI-driven location intelligence platform, crafting production-ready solutions that transform how businesses utilize and interpret location data. You will manage the full-stack development of our conversational AI system, encompassing LLM integration, agentic workflows, and real-time streaming interfaces, while laying the groundwork for our next-generation business intelligence tools. If you are a technical leader with over 10 years of software development experience, proven achievements in production Generative AI environments, and a zeal for constructing scalable AI platforms that yield tangible business results, we invite you to apply. RESPONSIBILITIES: Lead the technical design and architectural decisions for our GenAI products, ensuring robust scalability and performance of the AI-enabled location intelligence system. Develop technical roadmaps that align with business goals and scalability needs. Design and deploy comprehensive AI systems from prototype to production, including LLM integration, agentic architectures, and RAG implementations. Create resilient data pipelines and infrastructure capable of supporting AI/ML workloads at scale. Manage model deployment and integration processes, ensuring high availability and performance.
Join Credera, a pioneering global consulting firm that integrates transformative consulting skills, extensive industry expertise, AI, and technical proficiency to enhance customer experiences and drive growth across diverse sectors. We are constantly adapting our services to align with the future needs of organizations and adhere to contemporary best practices. Our distinct global approach offers customized solutions, reshaping some of the most impactful brands and entities worldwide.Our workforce, the core of our organization, is driven by a passion for making a significant difference in the lives of our clients, colleagues, and communities. This enthusiasm influences how we allocate our time, resources, and talents. We are proud of our commitment to our people and the recognition we've received globally. Visit our employer awards page for more details: https://www.credera.com/awards-and-recognition
Since its inception in 2007, Airbnb has transformed the way people experience travel, connecting over 5 million hosts with more than 2 billion guests across the globe. Our platform empowers hosts to offer memorable stays and unique experiences, fostering authentic connections within communities worldwide.Join Our Community:At Airbnb, delivering exceptional customer service is at the core of our mission. The complexity and scale of this endeavor present unique challenges and opportunities. Our Automation Foundation team within Community Support plays a pivotal role in shaping the ultimate customer service experience for our users.Your Impact:As a Senior Software Engineer, you will spearhead two essential initiatives within Airbnb's community support:Automation Platform: Lead our extensive conversational AI platform, collaborating with AI and ML engineering teams and cross-functional business partners to drive success in AI-driven products that benefit our community.Enhance the automation provisions for our internal human agents and AI agents.A Day in the Life: Work collaboratively with product, design, engineering, and data science teams to develop backend systems that optimize AI prompt effectiveness.Shape the technical vision and strategy for workflow and backend optimization.Lead the entire development cycle, including technical design, implementation, testing, experimentation, and deployment.Engage in code and design reviews, providing constructive feedback to elevate technical quality.Focus on optimizing system availability, performance, and scalability while pushing for innovative solutions.
Join pareto-ai as a Solutions Engineer and become a key player in our mission to deliver exceptional AI-driven solutions. In this fully remote role, you will collaborate with clients to understand their unique needs and provide tailored solutions that enhance their operational efficiency. With a focus on innovation and client satisfaction, you will utilize your technical expertise to troubleshoot issues and ensure seamless implementation of our products.
Join the Engineering Team at TRACTIANAt TRACTIAN, our Engineering team is dedicated to pioneering innovative infrastructure, technologies, and products that leverage the immense potential of IoT data. Our skilled engineers work collaboratively to create resilient systems, forward-thinking solutions, and scalable platforms that are key to TRACTIAN's success. We significantly influence the company’s strategic decisions, enhance operational efficiency, and ensure our customers receive outstanding experiences.Your RoleAs a GenAI Lead Engineer, you will be a vital member of the Software Team, partnering closely with developers, engineers, and product teams to orchestrate project organization, scheduling, and successful deployments. Your leadership skills will facilitate seamless coordination across various departments, ensuring objectives are met and tasks are completed proficiently. This role encompasses managing multiple components and provides the opportunity to engage in various facets of the Software Team's endeavors. You will play a crucial role in refining processes, boosting team efficiency, and directly impacting the overall success of the company.Key ResponsibilitiesProject Leadership: Strategically plan, schedule, and execute software projects on time and to the highest standards.Team Development: Oversee and mentor developers, promoting collaboration and high performance.Strategic Direction: Make critical decisions regarding backend architecture, scalability, and infrastructure.Cross-Functional Collaboration: Align software, product, engineering, and departmental objectives for flawless execution.Process Optimization: Streamline workflows, tackle technical challenges, and enhance productivity.Deployment Oversight: Ensure efficient deployments and maintain optimal system performance.Qualifications8+ years of experience in backend development with a focus on scalable system design, API architecture (DDD), and microservices (gRPC, REST), alongside expertise in databases (Postgres, MongoDB, Redis, Clickhouse) and cloud infrastructure (AWS).Proficient in TypeScript and Golang, with substantial experience in testing methodologies (unit, end-to-end, and load testing using Cypress, Playwright, k6).Skilled in containerization technologies (Docker, Kubernetes) and DevOps practices, including CI/CD pipelines, infrastructure as code (IaC), and system monitoring (Datadog, Grafana).Experience in developing and optimizing LLM applications, including Retrieval-Augmented Generation (RAG) and GPT-based solutions, with a strong foundation in deployment patterns, prompt engineering, and API integration (OpenAI, Anthropic, Gemini).Proficient in machine learning frameworks (PyTorch) and optimization techniques (quantization, model distillation).
Fully remote | Full-time engagement opportunityEstablished in Palo Alto by Dr. Andrew Ng and Israel Niezen, Factored is dedicated to empowering U.S. companies in building and scaling top-notch AI, ML, and Data teams, leveraging the exceptional talent from the top 1% of LATAM professionals. Our mission is clear: to empower brilliant minds, unleash their potential, and maximize their impact globally.At Factored, you will join a vibrant community that fosters learning, ownership, and authenticity. Your personal growth is prioritized, and your ideas are valued. Our culture is rooted in transparency, curiosity, and collaboration. We aim for excellence, celebrate diversity, promote curiosity, and create an environment where you can truly thrive.About Our Program:The Factored AI Machine Learning & GenAI Residency is a full-time, paid professional residency tailored for seasoned engineers aspiring to elevate their expertise and responsibility in impactful roles. Unlike conventional training programs, this residency offers a selective, execution-driven environment where engineers can validate and enhance their skills in developing and managing real-world Machine Learning and Generative AI systems in production.Residents will operate under genuine constraints, collaborating with peers who meet the same engineering standards, focusing on systems that truly matter.Program HighlightsA full-time, paid residency with defined professional expectations.Work on real production systems, moving beyond academic exercises.Take ownership of system-level decisions regarding architecture, performance, reliability, and cost.Engage in continuous technical evaluation through real execution rather than exams.Gain exposure to the standards for building and operating AI systems for U.S. companies.This residency provides a professional validation layer for senior engineers.Structure & SupportThe residency is structured as a professional track focused on execution.Residency Duration: 12 to 16 weeks, depending on individual readiness and scope.Time Commitment: Full-time (approximately 45 hours per week).Environment: Hands-on execution, technical reviews, and system-level discussions.
Full-time|$190.9K/yr - $232.8K/yr|On-site|San Francisco, California
P-1285 About This Role Join Databricks as a Staff Software Engineer specializing in GenAI inference, where you will spearhead the architecture, development, and optimization of the inference engine that powers the Databricks Foundation Model API. Your role will be crucial in bridging cutting-edge research with real-world production requirements, ensuring exceptional throughput, minimal latency, and scalable solutions. You will work across the entire GenAI inference stack, including kernels, runtimes, orchestration, memory management, and integration with various frameworks and orchestration systems. What You Will Do Take full ownership of the architecture, design, and implementation of the inference engine, collaborating on a model-serving stack optimized for large-scale LLM inference. Work closely with researchers to integrate new model architectures or features, such as sparsity, activation compression, and mixture-of-experts into the engine. Lead comprehensive optimization efforts focused on latency, throughput, memory efficiency, and hardware utilization across GPUs and other accelerators. Establish and uphold standards for building and maintaining instrumentation, profiling, and tracing tools to identify performance bottlenecks and drive optimizations. Design scalable solutions for routing, batching, scheduling, memory management, and dynamic loading tailored to inference workloads. Guarantee reliability, reproducibility, and fault tolerance in inference pipelines, including capabilities for A/B testing, rollbacks, and model versioning. Collaborate cross-functionally to integrate with federated and distributed inference infrastructure, ensuring effective orchestration across nodes, load balancing, and minimizing communication overhead. Foster collaboration with cross-functional teams, including platform engineers, cloud infrastructure, and security/compliance professionals. Represent the team externally through benchmarks, whitepapers, and contributions to open-source projects. What We Look For A BS/MS/PhD in Computer Science or a related discipline. A solid software engineering background with 6+ years of experience in performance-critical systems. A proven ability to own complex system components and influence architectural decisions from conception to execution. A deep understanding of ML inference internals, including attention mechanisms, MLPs, recurrent modules, quantization, and sparse operations. Hands-on experience with CUDA, GPU programming, and essential libraries (cuBLAS, cuDNN, NCCL, etc.). A strong foundation in distributed systems design, including RPC frameworks, queuing, RPC batching, sharding, and memory partitioning. Demonstrated proficiency in diagnosing and resolving performance bottlenecks across multiple layers (kernel, memory, networking, scheduler).
About Us:At novita-ai, we are a rapidly growing global provider of AI cloud infrastructure, leading the charge in the artificial intelligence revolution. Our innovative platform equips developers and enterprises with powerful, scalable, and user-friendly solutions such as Model APIs, GPU Instances, and Serverless Computing. As organizations around the globe strive to integrate AI into their offerings, we serve as the essential engine that fuels their innovative efforts.Join our world-class team and contribute to our expanding customer base. This unique opportunity allows you to be part of a dynamic company in a hyper-growth market, where your technical skills will directly impact customer success and drive our business forward.The Role:As a Solutions Engineer, you will act as the primary technical leader and trusted advisor for our clients throughout their journey. You will collaborate closely with the sales team to bridge the gap between complex customer challenges and our sophisticated technical solutions. Your mission is to build technical credibility, demonstrate the capabilities of our platform, and design tailored solutions that empower our clients to achieve their AI-related business objectives.What You'll Do:Technical Discovery & Solution Design: Collaborate with Account Executives to gain a deep understanding of customer needs, technical requirements, and business goals. Develop elegant and effective solutions utilizing our AI infrastructure stack (Model APIs, GPU Instances, Serverless).Product Demonstration & Proof of Concept (POC): Conduct engaging, customized product demonstrations and interactive workshops. Plan, manage, and execute successful POCs, showcasing the value and performance of our platform within the client’s environment.Technical Evangelism & Trusted Advisory: Communicate the value proposition of our platform to diverse audiences, including both technical and non-technical stakeholders, from engineers to C-level executives. Establish yourself as the go-to expert for customers on best practices in AI infrastructure.Sales Enablement & Market Feedback Loop: Create and maintain technical sales materials, including whitepapers, best practice guides, and demo scripts. Serve as the voice of the customer, relaying valuable feedback from the field to our Product and Engineering teams to influence our product roadmap.Onboarding & Implementation Guidance: Facilitate a seamless post-sales transition by providing initial onboarding support and architectural guidance, setting customers up for sustained success.
Join reddit as a Staff Machine Learning Engineer on our innovative GenAI Platform team. In this remote role, you will spearhead the development of cutting-edge AI solutions that enhance user experience and engagement across our platform. Your expertise in machine learning will help us unlock new capabilities and drive impactful changes in our product.
basis-ai develops intelligent agents that support complex operations for major accounting firms. With a recent $100 million funding round and a valuation above $1 billion, the company is backed by investors such as Khosla Ventures, Accel, Google Ventures, and leaders in technology. The team’s mission centers on bringing reliable, production-ready AI to industries where accuracy and dependability are essential. As a Solutions Engineer on the Deployed Intelligence team, the work centers on helping customers integrate basis-ai’s products into their daily operations. This includes handling technical integrations, overseeing key system connections, and translating customer requirements for the engineering team. Daily collaboration with DI Strategists is a core part of the role. Location This position is based in person at the Flatiron office in New York City. Preferred Qualifications Minimum 4 years of experience in solutions engineering, technical implementation, or a similar customer-facing technical field Strong skills in SQL for data analysis and troubleshooting Practical experience with API integrations, ideally within financial services or enterprise SaaS Proven problem-solving skills and a customer-focused approach
Point72 Asset Management is seeking a Machine Learning Infrastructure Engineer to help drive GenAI technology initiatives in the United States. This role centers on building and refining the systems that enable advanced AI projects across the firm. Key responsibilities Design and develop infrastructure specifically for GenAI applications Collaborate with data scientists and software engineers to create scalable and efficient solutions Optimize current systems to meet the changing requirements of AI models Role focus This position supports the foundation for GenAI projects by ensuring systems remain reliable and adaptable as technology evolves. Work involves both new development and ongoing improvements to existing infrastructure.
DescriptionWe are on the lookout for a pioneering Senior Applied AI Engineer who is not only adept at implementing AI solutions but is passionate about innovating them. This exciting opportunity is tailored for individuals who see the current AI landscape in cybersecurity as just the beginning and are driven to explore uncharted territories. You are the kind of visionary who developed AI agents long before they became mainstream, possess strong views on prompt engineering, and have the capability to design systems that others have yet to conceive.As a key member embedded within cyber units, you will act as a catalyst for change, revolutionizing the way security professionals operate. You will confront challenges that lack straightforward solutions, develop systems capable of managing genuine operational complexities, and create AI integrations that consistently perform under intense conditions. This role is not about following established patterns; it’s about inventing new paradigms. We seek candidates with substantial real-world experience in deploying AI systems to production—individuals who can distinguish between flashy demos and operational realities and possess the technical expertise to overcome obstacles where others falter.
Full-time|$300K/yr - $300K/yr|On-site|San Francisco
ABOUT BASETENBaseten is at the forefront of AI innovation, empowering top-tier companies like Cursor, Notion, OpenEvidence, Abridge, Clay, Gamma, and Writer to achieve mission-critical AI implementations. By integrating cutting-edge AI research, adaptable infrastructure, and streamlined developer tools, we help organizations harness the power of advanced models and bring them into production. As we continue to expand rapidly, having recently secured our $300M Series E funding round from esteemed investors such as BOND, IVP, Spark Capital, Greylock, and Conviction, we invite you to join our team and contribute to the platform that engineers rely on to deliver innovative AI products.THE ROLEAs an AI Solutions Engineer at Baseten, you will collaborate closely with clients to design, develop, and implement high-performance production AI applications utilizing Baseten's platform. You will guide customers through the entire process, from initial exploration to successful deployment, effectively translating complex business objectives into robust, observable services that deliver clear metrics on quality, latency, and cost.This position is ideal for proactive engineers eager to gain insights into how modern enterprises scale their AI adoption. You will thrive in a multidisciplinary environment, working across product development, software engineering, performance optimization, and direct customer engagement.It’s important to note that this is a hands-on engineering position that involves coding and software development, while also encompassing elements of product management, technical customer success, and pre-sales solution engineering.EXAMPLE INITIATIVESExplore the innovative projects undertaken by our Forward Deployed Engineering team:Forward Deployed Engineering on the Frontier of AIThe Fastest, Most Accurate Whisper TranscriptionDeploy Production-Ready Model Servers from Docker ImagesDeploy Custom ComfyUI Workflows as APIs
Join Tribe AI as an AI Solutions ArchitectAt Tribe AI, we are dedicated to empowering enterprises to unlock the transformative potential of AI. Many large organizations aspire to integrate AI into their operations, yet often lack the necessary expertise. This is where we step in.As a pioneering AI-native services firm, we collaborate with enterprises to design and implement top-tier AI solutions that generate tangible business outcomes. Our close partnerships with industry leaders like OpenAI and Anthropic provide us with unparalleled insights into cutting-edge models and strategic roadmaps.Your RoleIn the capacity of an AI Solutions Architect, you will bridge the gap between state-of-the-art AI capabilities and the practical needs of enterprises. Working alongside our internal AI specialists and General Managers, you will assess client requirements, identify impactful AI use cases, and craft customized solutions that are ready for deployment.This position goes beyond mere demonstrations; you will engage in technical discovery, evaluate architectural decisions, and offer clarity in uncertain situations, often validating concepts through rapid prototyping. Collaborating with leading model labs like OpenAI and Anthropic, you will navigate constraints and validate methodologies to harness emerging AI capabilities effectively.Key Responsibilities:Conduct technical discovery sessions with clients to identify workflows, limitations, data realities, and success metrics, translating these into actionable technical plans.Design and implement comprehensive AI systems that integrate Generative AI and traditional machine learning when necessary, covering data ingestion, orchestration, evaluation, deployment, and user experience.Create and present rapid prototypes that translate abstract concepts into tangible solutions, mitigating risks associated with cost, latency, reliability, integration feasibility, and quality thresholds.Develop reference architectures and technical frameworks that showcase the potential of current and emerging AI technologies.Guide clients through the AI ecosystem, discussing capabilities, limitations, trade-offs, and operational considerations for deploying AI systems effectively over time.Collaborate with sales teams to articulate technical concepts into compelling value propositions, proposals, business cases, and presentations.Specify technical requirements and scalable architecture patterns for production-grade systems, focusing on security, governance, observability, evaluation, and reliability.
Full-time|$166K/yr - $225K/yr|On-site|San Francisco, California
At Databricks, we are dedicated to empowering data teams to tackle the world's most challenging problems, from detecting security threats to advancing cancer drug development. We achieve this by offering the premier data and AI platform, allowing our customers to concentrate on their mission-critical challenges. The Mosaic AI organization assists companies in developing AI models and systems utilizing their own data, employing technologies that range from training large language models (LLMs) from the ground up to employing advanced retrieval methods for enhanced generation. We pride ourselves on pushing the boundaries of science and operationalizing our innovations. Mosaic AI believes that a company’s AI models hold intrinsic value, akin to any other core intellectual property, and that superior AI models should be accessible to all. Job Overview As a research engineer in the Scaling team, you will stay abreast of the latest advancements in deep learning and pioneer new methodologies that surpass the current state of the art. You will collaborate with a diverse team of researchers and engineers, sharing insights and expertise. Most importantly, you will be passionate about our customers, striving to ensure their success in implementing cutting-edge LLMs and AI systems by translating our scientific knowledge into practical applications. Your Impact Enhance performance through innovative optimization techniques, including kernel fusion, mixed precision, memory layout optimization, tiling strategies, and tensorization tailored for training-specific patterns. Design, implement, and optimize high-performance GPU kernels for training workloads, including attention mechanisms, custom layers, gradient computations, and activation functions, specifically for NVIDIA architectures. Create and implement distributed training frameworks for large language models, incorporating parallelism strategies (data, tensor, pipeline, ZeRO-based) and optimized communication patterns for gradient synchronization and collective operations. Profile, debug, and optimize comprehensive training workflows to pinpoint and resolve performance bottlenecks, utilizing memory optimization techniques such as activation checkpointing, gradient sharding, and mixed precision training.
As the Head of Global Workplace GenAI at Experian, you will be at the forefront of integrating cutting-edge Generative AI technologies into our workplace strategies. Your role will involve spearheading innovative solutions that enhance employee experiences and operational efficiencies across our global offices.Key responsibilities include leading a cross-functional team to design and implement AI-driven tools, fostering a culture of continuous improvement, and ensuring that our workplace technology aligns with Experian’s mission to empower consumers and businesses.
Full-time|$125K/yr - $175K/yr|On-site|San Francisco
About Arize AIIn a world rapidly evolving through artificial intelligence, the need for robust systems to monitor, troubleshoot, and optimize AI technologies has never been greater. At Arize AI, we lead the charge as the premier observability and evaluation platform for AI and Agent Engineering, equipping AI engineers to deploy high-performing, reliable agents and applications. From initial prototypes to full-scale production, our Arize AX platform consolidates build, test, and run into one seamless workspace, enabling teams to deliver solutions with speed and assurance.As a Series C company, we are supported by esteemed investors, having raised over $135M in funding and boasting a rapidly expanding clientele of 150+ prominent enterprises and Fortune 500 companies. Esteemed clients such as Booking.com, Uber, Siemens, and PepsiCo depend on Arize to create AI systems that deliver tangible results.Join Our TeamOur engineering team is dedicated to constructing systems that engage with some of the most intricate software ever deployed in live environments. Comprising seasoned industry professionals, our team has experience in developing deep learning infrastructures, autonomous drones, ridesharing services, ad technology, and much more.We are on the lookout for a client-centric AI Solutions Engineer who possesses a strong entrepreneurial spirit to join our Solutions Engineering organization. You will serve as a trusted technical advisor to our clients, driving business value, providing expert guidance, and nurturing account growth. Your role will involve leading clients to effective solutions, often by educating new users on our product or consulting on best practices. You should be prepared for technical discussions with data scientists and engineers while also conveying the value of Arize in business conversations with directors and executives. Your aim will be to empower our clients to become successful advocates of Arize.
About Arize AIAs the world rapidly evolves through AI advancements, the demand for robust monitoring, troubleshooting, and optimization of AI systems is more crucial than ever. At Arize AI, we lead the charge as the premier observability and evaluation platform for AI and Agent Engineering. Our platform, Arize AX, empowers AI engineers to deploy highly effective and reliable agents and applications seamlessly from initial prototypes to scaled production, fostering accelerated and confident deployment.As a Series C startup, we are backed by top-tier investors with over $135M in funding and a swiftly expanding clientele that includes 150+ prominent enterprises and Fortune 500 companies. Industry leaders like Booking.com, Uber, Siemens, and PepsiCo rely on Arize to ensure their AI solutions deliver exceptional performance.The TeamOur engineering team is dedicated to building systems that engage with some of the most intricate software deployed in real-world production. Comprising industry veterans with extensive experience in deep learning infrastructure, autonomous technologies, ridesharing platforms, and more, we are on a mission to redefine AI engineering.We seek a customer-centric AI Solutions Engineer with entrepreneurial spirit to join our Solutions Engineering team. As a trusted technical advisor, you will drive substantial business value, nurture client relationships, and expand accounts. Your role will involve guiding clients to effective solutions, educating new users on our product, and consulting on industry best practices. You must excel in technical discussions with data scientists and engineers, while also articulating the business value of Arize to directors and executives. Your primary aim will be to empower our clients to achieve success and become enthusiastic advocates of Arize.
Join soum as an AI / GenAI Solutions Engineer and contribute to cutting-edge projects that leverage artificial intelligence and generative AI technologies. In this pivotal role, you will be responsible for designing, developing, and implementing innovative AI solutions that enhance our product offerings and drive customer satisfaction.As part of our dynamic …
ABOUT PLACER.AI: Placer.ai is revolutionizing the way organizations comprehend the physical world. Our state-of-the-art location analytics platform offers unparalleled insights into locations, markets, and consumer behavior. By empowering a diverse range of customers—from Fortune 500 corporations to local governments and nonprofits—Placer enables data-driven decision-making. What distinguishes us? We have developed the most advanced location intelligence platform available while upholding an unwavering commitment to privacy. This demonstrates that powerful analytics can coexist with responsible data practices. Our rapid growth reflects the market's enthusiasm: in just six years since our inception, we achieved $100 million in annual recurring revenue, reached unicorn status with a valuation exceeding $1 billion in 2022, and continue to expand rapidly as one of North America's fastest-growing tech companies. We are creating a market opportunity worth over $100 billion and are only just getting started. Recognized as one of Forbes America's Best Startup Employers and a Deloitte Technology Fast 500 company, we are cultivating a culture where innovation flourishes, collaboration is paramount, and every team member plays a vital role in transforming how the world perceives location. SUMMARY: We are on the lookout for a seasoned Principal Engineer with extensive expertise in Generative AI to spearhead our GenAI product solutions and development initiatives. In this pivotal role, you will drive the technical vision and architecture of our AI-driven location intelligence platform, developing production-ready solutions that reshape how businesses harness and interpret location data. You will manage the full-stack development of our conversational AI system, encompassing LLM integration, agentic workflows, and real-time streaming interfaces, while laying the technical groundwork for our upcoming business intelligence tools. If you are a technical leader with over 10 years of software development experience, a proven track record in production GenAI settings, and a fervor for creating scalable AI platforms that yield tangible business benefits, we want to connect with you.
ABOUT PLACER.AI: At Placer.ai, we are revolutionizing the way organizations perceive and analyze the physical world. Our cutting-edge location analytics platform offers unmatched insights into locations, market trends, and consumer behaviors. Placer enables thousands of clients—from Fortune 500 companies to local governments and nonprofits—to make informed, data-driven decisions. What differentiates us? We have developed the most advanced location intelligence platform available while upholding an unwavering commitment to privacy, demonstrating that impactful analytics and ethical data practices can indeed coexist. Our rapid growth is a testament to the market's demand: we reached $100 million in annual recurring revenue within just six years of our inception, achieved unicorn status with a valuation exceeding $1 billion in 2022, and are consistently expanding as one of North America's fastest-growing tech firms. We are on the brink of creating a market opportunity exceeding $100 billion, and we are just getting started. Proudly recognized as one of Forbes' Best Startup Employers and a Deloitte Technology Fast 500 company, we are fostering a culture where innovation flourishes, collaboration thrives, and every member plays a pivotal role in redefining the understanding of location. SUMMARY: We are in search of a highly skilled Principal Engineer with extensive expertise in Generative AI to spearhead our GenAI product solutions and development initiatives. In this pivotal role, you will be responsible for shaping the technical vision and architecture of our AI-driven location intelligence platform, crafting production-ready solutions that transform how businesses utilize and interpret location data. You will manage the full-stack development of our conversational AI system, encompassing LLM integration, agentic workflows, and real-time streaming interfaces, while laying the groundwork for our next-generation business intelligence tools. If you are a technical leader with over 10 years of software development experience, proven achievements in production Generative AI environments, and a zeal for constructing scalable AI platforms that yield tangible business results, we invite you to apply. RESPONSIBILITIES: Lead the technical design and architectural decisions for our GenAI products, ensuring robust scalability and performance of the AI-enabled location intelligence system. Develop technical roadmaps that align with business goals and scalability needs. Design and deploy comprehensive AI systems from prototype to production, including LLM integration, agentic architectures, and RAG implementations. Create resilient data pipelines and infrastructure capable of supporting AI/ML workloads at scale. Manage model deployment and integration processes, ensuring high availability and performance.
Join Credera, a pioneering global consulting firm that integrates transformative consulting skills, extensive industry expertise, AI, and technical proficiency to enhance customer experiences and drive growth across diverse sectors. We are constantly adapting our services to align with the future needs of organizations and adhere to contemporary best practices. Our distinct global approach offers customized solutions, reshaping some of the most impactful brands and entities worldwide.Our workforce, the core of our organization, is driven by a passion for making a significant difference in the lives of our clients, colleagues, and communities. This enthusiasm influences how we allocate our time, resources, and talents. We are proud of our commitment to our people and the recognition we've received globally. Visit our employer awards page for more details: https://www.credera.com/awards-and-recognition
Since its inception in 2007, Airbnb has transformed the way people experience travel, connecting over 5 million hosts with more than 2 billion guests across the globe. Our platform empowers hosts to offer memorable stays and unique experiences, fostering authentic connections within communities worldwide.Join Our Community:At Airbnb, delivering exceptional customer service is at the core of our mission. The complexity and scale of this endeavor present unique challenges and opportunities. Our Automation Foundation team within Community Support plays a pivotal role in shaping the ultimate customer service experience for our users.Your Impact:As a Senior Software Engineer, you will spearhead two essential initiatives within Airbnb's community support:Automation Platform: Lead our extensive conversational AI platform, collaborating with AI and ML engineering teams and cross-functional business partners to drive success in AI-driven products that benefit our community.Enhance the automation provisions for our internal human agents and AI agents.A Day in the Life: Work collaboratively with product, design, engineering, and data science teams to develop backend systems that optimize AI prompt effectiveness.Shape the technical vision and strategy for workflow and backend optimization.Lead the entire development cycle, including technical design, implementation, testing, experimentation, and deployment.Engage in code and design reviews, providing constructive feedback to elevate technical quality.Focus on optimizing system availability, performance, and scalability while pushing for innovative solutions.
Join pareto-ai as a Solutions Engineer and become a key player in our mission to deliver exceptional AI-driven solutions. In this fully remote role, you will collaborate with clients to understand their unique needs and provide tailored solutions that enhance their operational efficiency. With a focus on innovation and client satisfaction, you will utilize your technical expertise to troubleshoot issues and ensure seamless implementation of our products.
Join the Engineering Team at TRACTIANAt TRACTIAN, our Engineering team is dedicated to pioneering innovative infrastructure, technologies, and products that leverage the immense potential of IoT data. Our skilled engineers work collaboratively to create resilient systems, forward-thinking solutions, and scalable platforms that are key to TRACTIAN's success. We significantly influence the company’s strategic decisions, enhance operational efficiency, and ensure our customers receive outstanding experiences.Your RoleAs a GenAI Lead Engineer, you will be a vital member of the Software Team, partnering closely with developers, engineers, and product teams to orchestrate project organization, scheduling, and successful deployments. Your leadership skills will facilitate seamless coordination across various departments, ensuring objectives are met and tasks are completed proficiently. This role encompasses managing multiple components and provides the opportunity to engage in various facets of the Software Team's endeavors. You will play a crucial role in refining processes, boosting team efficiency, and directly impacting the overall success of the company.Key ResponsibilitiesProject Leadership: Strategically plan, schedule, and execute software projects on time and to the highest standards.Team Development: Oversee and mentor developers, promoting collaboration and high performance.Strategic Direction: Make critical decisions regarding backend architecture, scalability, and infrastructure.Cross-Functional Collaboration: Align software, product, engineering, and departmental objectives for flawless execution.Process Optimization: Streamline workflows, tackle technical challenges, and enhance productivity.Deployment Oversight: Ensure efficient deployments and maintain optimal system performance.Qualifications8+ years of experience in backend development with a focus on scalable system design, API architecture (DDD), and microservices (gRPC, REST), alongside expertise in databases (Postgres, MongoDB, Redis, Clickhouse) and cloud infrastructure (AWS).Proficient in TypeScript and Golang, with substantial experience in testing methodologies (unit, end-to-end, and load testing using Cypress, Playwright, k6).Skilled in containerization technologies (Docker, Kubernetes) and DevOps practices, including CI/CD pipelines, infrastructure as code (IaC), and system monitoring (Datadog, Grafana).Experience in developing and optimizing LLM applications, including Retrieval-Augmented Generation (RAG) and GPT-based solutions, with a strong foundation in deployment patterns, prompt engineering, and API integration (OpenAI, Anthropic, Gemini).Proficient in machine learning frameworks (PyTorch) and optimization techniques (quantization, model distillation).
Fully remote | Full-time engagement opportunityEstablished in Palo Alto by Dr. Andrew Ng and Israel Niezen, Factored is dedicated to empowering U.S. companies in building and scaling top-notch AI, ML, and Data teams, leveraging the exceptional talent from the top 1% of LATAM professionals. Our mission is clear: to empower brilliant minds, unleash their potential, and maximize their impact globally.At Factored, you will join a vibrant community that fosters learning, ownership, and authenticity. Your personal growth is prioritized, and your ideas are valued. Our culture is rooted in transparency, curiosity, and collaboration. We aim for excellence, celebrate diversity, promote curiosity, and create an environment where you can truly thrive.About Our Program:The Factored AI Machine Learning & GenAI Residency is a full-time, paid professional residency tailored for seasoned engineers aspiring to elevate their expertise and responsibility in impactful roles. Unlike conventional training programs, this residency offers a selective, execution-driven environment where engineers can validate and enhance their skills in developing and managing real-world Machine Learning and Generative AI systems in production.Residents will operate under genuine constraints, collaborating with peers who meet the same engineering standards, focusing on systems that truly matter.Program HighlightsA full-time, paid residency with defined professional expectations.Work on real production systems, moving beyond academic exercises.Take ownership of system-level decisions regarding architecture, performance, reliability, and cost.Engage in continuous technical evaluation through real execution rather than exams.Gain exposure to the standards for building and operating AI systems for U.S. companies.This residency provides a professional validation layer for senior engineers.Structure & SupportThe residency is structured as a professional track focused on execution.Residency Duration: 12 to 16 weeks, depending on individual readiness and scope.Time Commitment: Full-time (approximately 45 hours per week).Environment: Hands-on execution, technical reviews, and system-level discussions.
Full-time|$190.9K/yr - $232.8K/yr|On-site|San Francisco, California
P-1285 About This Role Join Databricks as a Staff Software Engineer specializing in GenAI inference, where you will spearhead the architecture, development, and optimization of the inference engine that powers the Databricks Foundation Model API. Your role will be crucial in bridging cutting-edge research with real-world production requirements, ensuring exceptional throughput, minimal latency, and scalable solutions. You will work across the entire GenAI inference stack, including kernels, runtimes, orchestration, memory management, and integration with various frameworks and orchestration systems. What You Will Do Take full ownership of the architecture, design, and implementation of the inference engine, collaborating on a model-serving stack optimized for large-scale LLM inference. Work closely with researchers to integrate new model architectures or features, such as sparsity, activation compression, and mixture-of-experts into the engine. Lead comprehensive optimization efforts focused on latency, throughput, memory efficiency, and hardware utilization across GPUs and other accelerators. Establish and uphold standards for building and maintaining instrumentation, profiling, and tracing tools to identify performance bottlenecks and drive optimizations. Design scalable solutions for routing, batching, scheduling, memory management, and dynamic loading tailored to inference workloads. Guarantee reliability, reproducibility, and fault tolerance in inference pipelines, including capabilities for A/B testing, rollbacks, and model versioning. Collaborate cross-functionally to integrate with federated and distributed inference infrastructure, ensuring effective orchestration across nodes, load balancing, and minimizing communication overhead. Foster collaboration with cross-functional teams, including platform engineers, cloud infrastructure, and security/compliance professionals. Represent the team externally through benchmarks, whitepapers, and contributions to open-source projects. What We Look For A BS/MS/PhD in Computer Science or a related discipline. A solid software engineering background with 6+ years of experience in performance-critical systems. A proven ability to own complex system components and influence architectural decisions from conception to execution. A deep understanding of ML inference internals, including attention mechanisms, MLPs, recurrent modules, quantization, and sparse operations. Hands-on experience with CUDA, GPU programming, and essential libraries (cuBLAS, cuDNN, NCCL, etc.). A strong foundation in distributed systems design, including RPC frameworks, queuing, RPC batching, sharding, and memory partitioning. Demonstrated proficiency in diagnosing and resolving performance bottlenecks across multiple layers (kernel, memory, networking, scheduler).
About Us:At novita-ai, we are a rapidly growing global provider of AI cloud infrastructure, leading the charge in the artificial intelligence revolution. Our innovative platform equips developers and enterprises with powerful, scalable, and user-friendly solutions such as Model APIs, GPU Instances, and Serverless Computing. As organizations around the globe strive to integrate AI into their offerings, we serve as the essential engine that fuels their innovative efforts.Join our world-class team and contribute to our expanding customer base. This unique opportunity allows you to be part of a dynamic company in a hyper-growth market, where your technical skills will directly impact customer success and drive our business forward.The Role:As a Solutions Engineer, you will act as the primary technical leader and trusted advisor for our clients throughout their journey. You will collaborate closely with the sales team to bridge the gap between complex customer challenges and our sophisticated technical solutions. Your mission is to build technical credibility, demonstrate the capabilities of our platform, and design tailored solutions that empower our clients to achieve their AI-related business objectives.What You'll Do:Technical Discovery & Solution Design: Collaborate with Account Executives to gain a deep understanding of customer needs, technical requirements, and business goals. Develop elegant and effective solutions utilizing our AI infrastructure stack (Model APIs, GPU Instances, Serverless).Product Demonstration & Proof of Concept (POC): Conduct engaging, customized product demonstrations and interactive workshops. Plan, manage, and execute successful POCs, showcasing the value and performance of our platform within the client’s environment.Technical Evangelism & Trusted Advisory: Communicate the value proposition of our platform to diverse audiences, including both technical and non-technical stakeholders, from engineers to C-level executives. Establish yourself as the go-to expert for customers on best practices in AI infrastructure.Sales Enablement & Market Feedback Loop: Create and maintain technical sales materials, including whitepapers, best practice guides, and demo scripts. Serve as the voice of the customer, relaying valuable feedback from the field to our Product and Engineering teams to influence our product roadmap.Onboarding & Implementation Guidance: Facilitate a seamless post-sales transition by providing initial onboarding support and architectural guidance, setting customers up for sustained success.
Join reddit as a Staff Machine Learning Engineer on our innovative GenAI Platform team. In this remote role, you will spearhead the development of cutting-edge AI solutions that enhance user experience and engagement across our platform. Your expertise in machine learning will help us unlock new capabilities and drive impactful changes in our product.
basis-ai develops intelligent agents that support complex operations for major accounting firms. With a recent $100 million funding round and a valuation above $1 billion, the company is backed by investors such as Khosla Ventures, Accel, Google Ventures, and leaders in technology. The team’s mission centers on bringing reliable, production-ready AI to industries where accuracy and dependability are essential. As a Solutions Engineer on the Deployed Intelligence team, the work centers on helping customers integrate basis-ai’s products into their daily operations. This includes handling technical integrations, overseeing key system connections, and translating customer requirements for the engineering team. Daily collaboration with DI Strategists is a core part of the role. Location This position is based in person at the Flatiron office in New York City. Preferred Qualifications Minimum 4 years of experience in solutions engineering, technical implementation, or a similar customer-facing technical field Strong skills in SQL for data analysis and troubleshooting Practical experience with API integrations, ideally within financial services or enterprise SaaS Proven problem-solving skills and a customer-focused approach
Point72 Asset Management is seeking a Machine Learning Infrastructure Engineer to help drive GenAI technology initiatives in the United States. This role centers on building and refining the systems that enable advanced AI projects across the firm. Key responsibilities Design and develop infrastructure specifically for GenAI applications Collaborate with data scientists and software engineers to create scalable and efficient solutions Optimize current systems to meet the changing requirements of AI models Role focus This position supports the foundation for GenAI projects by ensuring systems remain reliable and adaptable as technology evolves. Work involves both new development and ongoing improvements to existing infrastructure.
DescriptionWe are on the lookout for a pioneering Senior Applied AI Engineer who is not only adept at implementing AI solutions but is passionate about innovating them. This exciting opportunity is tailored for individuals who see the current AI landscape in cybersecurity as just the beginning and are driven to explore uncharted territories. You are the kind of visionary who developed AI agents long before they became mainstream, possess strong views on prompt engineering, and have the capability to design systems that others have yet to conceive.As a key member embedded within cyber units, you will act as a catalyst for change, revolutionizing the way security professionals operate. You will confront challenges that lack straightforward solutions, develop systems capable of managing genuine operational complexities, and create AI integrations that consistently perform under intense conditions. This role is not about following established patterns; it’s about inventing new paradigms. We seek candidates with substantial real-world experience in deploying AI systems to production—individuals who can distinguish between flashy demos and operational realities and possess the technical expertise to overcome obstacles where others falter.
Full-time|$300K/yr - $300K/yr|On-site|San Francisco
ABOUT BASETENBaseten is at the forefront of AI innovation, empowering top-tier companies like Cursor, Notion, OpenEvidence, Abridge, Clay, Gamma, and Writer to achieve mission-critical AI implementations. By integrating cutting-edge AI research, adaptable infrastructure, and streamlined developer tools, we help organizations harness the power of advanced models and bring them into production. As we continue to expand rapidly, having recently secured our $300M Series E funding round from esteemed investors such as BOND, IVP, Spark Capital, Greylock, and Conviction, we invite you to join our team and contribute to the platform that engineers rely on to deliver innovative AI products.THE ROLEAs an AI Solutions Engineer at Baseten, you will collaborate closely with clients to design, develop, and implement high-performance production AI applications utilizing Baseten's platform. You will guide customers through the entire process, from initial exploration to successful deployment, effectively translating complex business objectives into robust, observable services that deliver clear metrics on quality, latency, and cost.This position is ideal for proactive engineers eager to gain insights into how modern enterprises scale their AI adoption. You will thrive in a multidisciplinary environment, working across product development, software engineering, performance optimization, and direct customer engagement.It’s important to note that this is a hands-on engineering position that involves coding and software development, while also encompassing elements of product management, technical customer success, and pre-sales solution engineering.EXAMPLE INITIATIVESExplore the innovative projects undertaken by our Forward Deployed Engineering team:Forward Deployed Engineering on the Frontier of AIThe Fastest, Most Accurate Whisper TranscriptionDeploy Production-Ready Model Servers from Docker ImagesDeploy Custom ComfyUI Workflows as APIs
Join Tribe AI as an AI Solutions ArchitectAt Tribe AI, we are dedicated to empowering enterprises to unlock the transformative potential of AI. Many large organizations aspire to integrate AI into their operations, yet often lack the necessary expertise. This is where we step in.As a pioneering AI-native services firm, we collaborate with enterprises to design and implement top-tier AI solutions that generate tangible business outcomes. Our close partnerships with industry leaders like OpenAI and Anthropic provide us with unparalleled insights into cutting-edge models and strategic roadmaps.Your RoleIn the capacity of an AI Solutions Architect, you will bridge the gap between state-of-the-art AI capabilities and the practical needs of enterprises. Working alongside our internal AI specialists and General Managers, you will assess client requirements, identify impactful AI use cases, and craft customized solutions that are ready for deployment.This position goes beyond mere demonstrations; you will engage in technical discovery, evaluate architectural decisions, and offer clarity in uncertain situations, often validating concepts through rapid prototyping. Collaborating with leading model labs like OpenAI and Anthropic, you will navigate constraints and validate methodologies to harness emerging AI capabilities effectively.Key Responsibilities:Conduct technical discovery sessions with clients to identify workflows, limitations, data realities, and success metrics, translating these into actionable technical plans.Design and implement comprehensive AI systems that integrate Generative AI and traditional machine learning when necessary, covering data ingestion, orchestration, evaluation, deployment, and user experience.Create and present rapid prototypes that translate abstract concepts into tangible solutions, mitigating risks associated with cost, latency, reliability, integration feasibility, and quality thresholds.Develop reference architectures and technical frameworks that showcase the potential of current and emerging AI technologies.Guide clients through the AI ecosystem, discussing capabilities, limitations, trade-offs, and operational considerations for deploying AI systems effectively over time.Collaborate with sales teams to articulate technical concepts into compelling value propositions, proposals, business cases, and presentations.Specify technical requirements and scalable architecture patterns for production-grade systems, focusing on security, governance, observability, evaluation, and reliability.
Full-time|$166K/yr - $225K/yr|On-site|San Francisco, California
At Databricks, we are dedicated to empowering data teams to tackle the world's most challenging problems, from detecting security threats to advancing cancer drug development. We achieve this by offering the premier data and AI platform, allowing our customers to concentrate on their mission-critical challenges. The Mosaic AI organization assists companies in developing AI models and systems utilizing their own data, employing technologies that range from training large language models (LLMs) from the ground up to employing advanced retrieval methods for enhanced generation. We pride ourselves on pushing the boundaries of science and operationalizing our innovations. Mosaic AI believes that a company’s AI models hold intrinsic value, akin to any other core intellectual property, and that superior AI models should be accessible to all. Job Overview As a research engineer in the Scaling team, you will stay abreast of the latest advancements in deep learning and pioneer new methodologies that surpass the current state of the art. You will collaborate with a diverse team of researchers and engineers, sharing insights and expertise. Most importantly, you will be passionate about our customers, striving to ensure their success in implementing cutting-edge LLMs and AI systems by translating our scientific knowledge into practical applications. Your Impact Enhance performance through innovative optimization techniques, including kernel fusion, mixed precision, memory layout optimization, tiling strategies, and tensorization tailored for training-specific patterns. Design, implement, and optimize high-performance GPU kernels for training workloads, including attention mechanisms, custom layers, gradient computations, and activation functions, specifically for NVIDIA architectures. Create and implement distributed training frameworks for large language models, incorporating parallelism strategies (data, tensor, pipeline, ZeRO-based) and optimized communication patterns for gradient synchronization and collective operations. Profile, debug, and optimize comprehensive training workflows to pinpoint and resolve performance bottlenecks, utilizing memory optimization techniques such as activation checkpointing, gradient sharding, and mixed precision training.
As the Head of Global Workplace GenAI at Experian, you will be at the forefront of integrating cutting-edge Generative AI technologies into our workplace strategies. Your role will involve spearheading innovative solutions that enhance employee experiences and operational efficiencies across our global offices.Key responsibilities include leading a cross-functional team to design and implement AI-driven tools, fostering a culture of continuous improvement, and ensuring that our workplace technology aligns with Experian’s mission to empower consumers and businesses.
Full-time|$125K/yr - $175K/yr|On-site|San Francisco
About Arize AIIn a world rapidly evolving through artificial intelligence, the need for robust systems to monitor, troubleshoot, and optimize AI technologies has never been greater. At Arize AI, we lead the charge as the premier observability and evaluation platform for AI and Agent Engineering, equipping AI engineers to deploy high-performing, reliable agents and applications. From initial prototypes to full-scale production, our Arize AX platform consolidates build, test, and run into one seamless workspace, enabling teams to deliver solutions with speed and assurance.As a Series C company, we are supported by esteemed investors, having raised over $135M in funding and boasting a rapidly expanding clientele of 150+ prominent enterprises and Fortune 500 companies. Esteemed clients such as Booking.com, Uber, Siemens, and PepsiCo depend on Arize to create AI systems that deliver tangible results.Join Our TeamOur engineering team is dedicated to constructing systems that engage with some of the most intricate software ever deployed in live environments. Comprising seasoned industry professionals, our team has experience in developing deep learning infrastructures, autonomous drones, ridesharing services, ad technology, and much more.We are on the lookout for a client-centric AI Solutions Engineer who possesses a strong entrepreneurial spirit to join our Solutions Engineering organization. You will serve as a trusted technical advisor to our clients, driving business value, providing expert guidance, and nurturing account growth. Your role will involve leading clients to effective solutions, often by educating new users on our product or consulting on best practices. You should be prepared for technical discussions with data scientists and engineers while also conveying the value of Arize in business conversations with directors and executives. Your aim will be to empower our clients to become successful advocates of Arize.
About Arize AIAs the world rapidly evolves through AI advancements, the demand for robust monitoring, troubleshooting, and optimization of AI systems is more crucial than ever. At Arize AI, we lead the charge as the premier observability and evaluation platform for AI and Agent Engineering. Our platform, Arize AX, empowers AI engineers to deploy highly effective and reliable agents and applications seamlessly from initial prototypes to scaled production, fostering accelerated and confident deployment.As a Series C startup, we are backed by top-tier investors with over $135M in funding and a swiftly expanding clientele that includes 150+ prominent enterprises and Fortune 500 companies. Industry leaders like Booking.com, Uber, Siemens, and PepsiCo rely on Arize to ensure their AI solutions deliver exceptional performance.The TeamOur engineering team is dedicated to building systems that engage with some of the most intricate software deployed in real-world production. Comprising industry veterans with extensive experience in deep learning infrastructure, autonomous technologies, ridesharing platforms, and more, we are on a mission to redefine AI engineering.We seek a customer-centric AI Solutions Engineer with entrepreneurial spirit to join our Solutions Engineering team. As a trusted technical advisor, you will drive substantial business value, nurture client relationships, and expand accounts. Your role will involve guiding clients to effective solutions, educating new users on our product, and consulting on industry best practices. You must excel in technical discussions with data scientists and engineers, while also articulating the business value of Arize to directors and executives. Your primary aim will be to empower our clients to achieve success and become enthusiastic advocates of Arize.
Feb 2, 2026
Sign in to browse more jobs
Create account — see all 56,970 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.