Openai Jobs in Usa
75 positions found — Page 4
Research Scientist, Post-Training | Robotics & AI | SF Bay Area / Boston
A frontier AI robotics company building general-purpose robots is hiring a Post-Training Research Scientist to turn large pretrained robot models into reliable real-world systems. The team includes alumni from OpenAI, Boston Dynamics, and Google DeepMind, working on embodied foundation models and robotic dexterity.
Role
Focus on post-training: fine-tuning and adapting large models for real robotic tasks, improving robustness and controllability, and building evaluation frameworks tied to real-world performance. Work on inference efficiency and close the loop between model outputs and robot behaviour using RL, imitation learning, distillation, synthetic data, and curriculum learning.
Looking for
Experience fine-tuning large models (RLHF, RL, IL, distillation, domain adaptation), ideally with exposure to embodied AI or real-world ML systems, plus strong evaluation and debugging skills.
On-site: SF Bay Area or Boston
Up to $350k + equity
Proper Hospitality is seeking an AI Workflow Fellow for a three month, execution focused program embedded with the CEO, President, and Chief of Staff. This role is responsible for building and deploying production ready AI workflows across our hotel portfolio, automating real operational processes tied to clear ROI, and integrating across systems including PMS, Snowflake, Microsoft 365, and guest experience platforms. This is hands on applied AI with live impact on property operations, not research or prototype work.
What This Is
We're not exploring AI at Proper Hotels. We're deploying it. Right now, AI runs our executive briefings, triages guest reviews across 11 properties, automates reporting pipelines, and handles operational workflows that used to eat hundreds of hours. We need someone who can build more of this, faster.
This is a single fellowship seat. You'll work directly with our CEO, President, Chief of Staff and executive team for three months and will be shipping production workflows from week one.
You are the execution engine inside Proper's broader Workflow & AI operating model.
What You'll Actually Do
Building (80%)
Build & Ship
- Design and deploy agent-based workflows that automate real operational tasks (not demos, not prototypes that sit on a shelf)
- Build within clearly defined ROI hypotheses approved by the Head of Workflow & AI (not speculative experiments)
- Wire together APIs across our stack: PMS, Snowflake, Notion, Microsoft 365, Google Workspace, Revinate, STR
- Build multi-agent systems that handle overnight operations, reporting rollups, task accountability, and guest experience triage
- Create tools that General Managers and department heads actually use daily
Identify & Automate
- Audit departmental workflows across the portfolio and find the manual processes burning the most hours
- Build the automation, test it on-property, iterate based on real feedback
- Transition tasks from \"someone does this by hand\" to \"this runs itself\" without losing the human touch that defines Proper
Strategy (20%)
Strategic Input
- Evaluate frontier capabilities weekly, but only deploy those that map to defined operational ROI
- Translate what's happening at the AI frontier into specific, actionable opportunities for luxury hospitality
- Help shape our internal AI skill-building program so the culture evolves with the technology
Who You Are
- You build agents and workflows, not just prompts. Show us something you've built that runs without you babysitting it
- You've shipped applied AI into production environments. Side projects count if they're real and running
- You can wire APIs together before lunch and present to the C-suite after it
- You navigate ambiguity without freezing. If a tool doesn't exist, you build it
- You understand that technology in a hotel should be invisible but felt. \"High Tech / High Touch\" isn't a slogan to you
- You're hands-on with LLMs (OpenAI, Anthropic, open-source), API orchestration, agent frameworks (eg. Openclaw), and data pipelines
- Bonus: experience with hospitality systems, revenue management, or guest experience platforms
Education
CS, Data Science, or MBA with a strong technical background preferred but not required. Non-traditional paths welcome if your portfolio speaks for itself
Program Details
- Duration: 3 months with potential to extend
- Experience: 0-2 years
- Compensation: $7,000 - $10,000/month depending on experience and location
- Access: Direct seat at the table with the CEO, President, and Chief of Staff
- Impact: Your work goes live on-property, affecting real guests and real revenue. This isn't a sandbox.
In your application please include two additional items:
- Something you've built that automates a real workflow (link, repo, or demo)
- A short note on what you'd build first if you had access to a luxury hotel portfolio's entire data stack
Why Join Proper Hospitality
At Proper, we build experiences that move people — and that begins with the team behind them. As a best-in-class employer, we're committed to creating one of the Best Places to Work in hospitality by nurturing a culture where creativity, excellence, and humanity thrive together.
Everything we do is grounded in the belief that hospitality is more than a profession - it's an opportunity to care for others and make lives better. Guided by the Pillars of Proper, we show up with warmth and authenticity (Care Proper), strive for excellence in everything we do (Achieve Proper), think creatively and resourcefully (Imagine Proper), and take pride in the style and culture that make us who we are (Present Proper).
We believe our people are our greatest strength, and we invest deeply in their wellbeing, growth, and sense of belonging. From comprehensive benefits to meaningful development programs, Proper is designed to help you build a career, and a life, that feels as inspiring as the experiences we create for our guests.
Our Commitment: Building the Best Place to Work
Our Best Place to Work initiative is a living commitment — a continuous investment in our people, our culture, and our purpose. We listen, learn, and evolve together to create an environment where everyone feels empowered to imagine boldly, achieve confidently, care deeply, and present themselves authentically.
At Proper, joining the team means more than finding a job — it means joining a community that believes in building beautiful experiences together, for our guests and for one another.
Title: Full Stack Developer with AI
Duration: 12 Months+
Location: Spring, TX
Type: Onsite
We are seeking a Full Stack Developer who will contribute to building scalable backend services including platform and utility modules application. You will also play an active role in implementing GenAI use cases using modern agentic frameworks.
You will collaborate with product owner, trading fusion developers, data engineers, and other full stack developers across regions.
Responsibilities?:?
- Platform Engineering & Support
- Develop, enhance, and support components of the Global Trading App platform
- Implement monitoring, alerting, and telemetry capabilities using modern observability tools
- Improve platform reliability, scalability, and performance through proactive engineering
- Author infrastructure-as-code using Terraform for cloud resources
Application & Service Development
- Build secure and scalable backend APIs (primarily in Python / FastAPI)
- Create responsive and efficient React-based UI components
- Develop reusable utility modules for fusion teams to accelerate delivery
GenAI & Agentic Solutions
- Implement GenAI-powered features using LLMs, vector databases, and multi-agent frameworks
- Develop "agentic" workflows for automation, troubleshooting, and developer productivity
- Build model integration and evaluation
Collaboration & Standards
- Contribute to engineering best practices and documentation
- Work closely with global trading fusion teams to ensure alignment and technical excellence
Qualifications?:
- Python (advanced): APIs, data processing, async programming
- React: modern component-based UI development
- FastAPI: building high performance backend services
- DBT: data engineering and transformation
- GitHub/CI/CD: strong version control and build pipeline experience
Preferred Skills:
- Terraform, Azure, AWS: infrastructure provisioning and automation
- Databricks, Snowflake
- GenAI / Multi-Agent
- Experience implementing solutions using LLMs, embeddings, prompt engineering
- Familiarity with agentic coding frameworks (e.g., LangChain, AutoGen, OpenAI agents, etc.)
- Understanding of RAG, model orchestration, and AI application patterns
Soft Skills:
- Strong problem-solving skills and ownership mindset
- Ability to work in global, cross-functional teams
- Clear communication and documentation abilities
- Comfort operating in fast-paced, high-availability environments
- Adaptability and willingness to learn new technologies and methodologies
This Jobot Job is hosted by: Charles Simmons
Are you a fit? Easy Apply now by clicking the "Apply" button
and sending us your resume.
Salary: $140,000 - $180,000 per year
A bit about us:
We’re transforming how government agencies digitize forms and automate workflows. Our new initiative brings AI directly into this process - using LLMs, vector search, and structured PDF parsing to accelerate public service delivery. We’re not just bolting AI onto the side. It’s becoming core to how our platform works.
We’re looking for a senior machine learning engineer to take the lead on this effort. You’ll be the architect of our AI capability - not just a contributor. Your work will touch thousands of public-facing government forms, helping real people get things done faster and more accurately. This isn’t an R&D team running experiments - it’s about delivering intelligent automation, right now.
Why join us?
- 100% remote based in the US
- Help shape the AI transformation of public sector services
- Lead initiatives that ship real impact, not just prototypes
- Greenfield development on a proven, profitable platform
- Comprehensive Health, Vision, Dental coverage for individuals and families
Job Details
You’ll design and build our machine learning infrastructure - starting with vector search and retrieval-augmented generation and expanding into fine-tuned LLMs with human feedback loops. You’ll work across product and engineering to embed intelligent behaviors into our no-code form builder. This is not a research job or a sandbox role - it’s a real opportunity to push AI into production at scale.
What you’ll do
- Build and tune vector-based retrieval pipelines using OpenAI embeddings and Azure AI Search
- Design prompt strategies and agents to translate parsed PDF data into form component schemas
- Fine-tune LLMs for structured output generation with low-latency performance in mind
- Lead the development of an RLHF loop that incorporates builder UI feedback and audit data
- Help architect systems that blend traditional APIs and probabilistic inference reliably
- Work alongside full-stack and platform engineers to get it all running in production
- Stay plugged into the latest model capabilities, and make smart calls on what to adopt
Tech you’ll use
- Azure AI Studio, Azure OpenAI, GPT-4o
- Python (for agents, functions, orchestration), .NET 8 (for integration layers)
- Azure AI Search, CosmosDB, MSSQL
- Kubernetes (AKS), Azure Blob, Octopus for CI/CD
- for structured PDF parsing
What we’re looking for
- 5+ years in applied ML, including experience with retrieval, embeddings, and prompt engineering
- Strong Python skills and familiarity with production-grade ML pipelines
- Experience designing and tuning RAG workflows with hybrid search
- Familiarity with RLHF and fine-tuning on structured JSON output
- Solid grasp of system-level thinking—how to bring ML into product environments cleanly
Nice to have: .NET understanding, especially for integration and orchestration layers
What success looks like in 6 months
- You’ve shipped a working vector search + RAG pipeline integrated into our form builder
- You’ve scoped and kicked off our first LLM fine-tuning cycle
- We’re collecting human feedback to improve model accuracy
- You’ve helped define the roadmap for AI integrations across the platform
Interested in hearing more? Easy Apply now by clicking the "Apply" button.
Jobot is an Equal Opportunity Employer. We provide an inclusive work environment that celebrates diversity and all qualified candidates receive consideration for employment without regard to race, color, sex, sexual orientation, gender identity, religion, national origin, age (40 and over), disability, military status, genetic information or any other basis protected by applicable federal, state, or local laws. Jobot also prohibits harassment of applicants or employees based on any of these protected categories. It is Jobot’s policy to comply with all applicable federal, state and local laws respecting consideration of unemployment status in making hiring decisions.
Sometimes Jobot is required to perform background checks with your authorization. Jobot will consider qualified candidates with criminal histories in a manner consistent with any applicable federal, state, or local law regarding criminal backgrounds, including but not limited to the Los Angeles Fair Chance Initiative for Hiring and the San Francisco Fair Chance Ordinance.
Information collected and processed as part of your Jobot candidate profile, and any job applications, resumes, or other information you choose to submit is subject to Jobot's Privacy Policy, as well as the Jobot California Worker Privacy Notice and Jobot Notice Regarding Automated Employment Decision Tools which are available at /legal.
By applying for this job, you agree to receive calls, AI-generated calls, text messages, or emails from Jobot, and/or its agents and contracted partners. Frequency varies for text messages. Message and data rates may apply. Carriers are not liable for delayed or undelivered messages. You can reply STOP to cancel and HELP for help. You can access our privacy policy here: /privacy-policy
The Opportunity
The Head of Enterprise Productivity & Collaboration Tools is a senior leader responsible for the strategic direction, integration, and optimization of MassMutual's core collaboration and productivity platforms—including ServiceNow, Microsoft 365, Slack, Zoom, Teams, and Atlassian—with a strong focus on AI enablement and automation. This role will drive the adoption of AI-powered capabilities across platforms to enhance employee experience, operational efficiency, and business agility.
The Team
You will lead a multi-disciplinary global team, including senior managers and specialized platform experts. The team is collaborative, agile, and focused on innovation and accountability. You will foster a high-performance, inclusive culture and develop talent strategies to build AI fluency and platform expertise across the organization.
The Impact:
- Develop and execute a 1–3+ year strategic roadmap for enterprise enablement platforms with a focus on AI integration, automation, and intelligent workflows
- Partner with Enterprise Architecture and Research & Development teams to embed AI capabilities into platform services (e.g., virtual agents, intelligent routing, predictive analytics)
- Lead platform modernization efforts, ensuring alignment with digital transformation and AI-first strategies
- Promote a culture of innovation by enabling employees to contribute ideas and participate in proof-of-concept initiatives
- Oversee platform governance, performance, and lifecycle management across enterprise platforms including ServiceNow, Microsoft, Slack, Zoom, Teams, and Atlassian
- Define and monitor KPIs, SLAs, and AI-driven metrics to ensure platform reliability, adoption, and value realization
- Optimize licensing, usage, and cost efficiency through data-driven insights and automation
- Serve as a strategic advisor to senior leadership on platform and AI enablement opportunities
- Collaborate with business, IT, security, and compliance stakeholders to ensure alignment and adoption
- Lead change management efforts to drive user engagement and maximize ROI from platform investments
- Manage, coach, and motivate a global team while building a strong, diverse talent pipeline
- Manage vendor relationships and external partnerships
- Execute operational strategy with latitude in day-to-day decision-making
- Implement short- to medium-term strategies (1–3 years); contribute to broader strategic planning
- Manage teams performing multiple functions/disciplines, including exempt and non-exempt employees
The Minimum Qualifications
- Bachelor's degree in Technology, Business, or related field
- 10+ years of experience in enterprise technology, with 5+ years in senior leadership roles
- Proven track record in managing and scaling enterprise platforms (ServiceNow, Microsoft Copilot, Slack, Zoom, Teams, Atlassian)
- Deep understanding of AI technologies and their application in enterprise environments
- Experience leading platform modernization and AI transformation initiatives
- Strong financial acumen with experience managing large-scale budgets
- Exceptional communication, stakeholder management, and strategic planning skills
- Authorized to work in the U.S.
The Ideal Qualifications
- Experience with AI/ML platforms (e.g., Azure AI, OpenAI, ServiceNow AI, Microsoft Copilot)
- Familiarity with agile, DevOps, and product management methodologies
- MBA or advanced degree in a relevant field
- Excellent facilitation, collaboration, and change management skills
- Strong financial acumen and data-driven decision-making
- Ability to work across all organizational levels and in complex environments
- Global travel flexibility and self-driven work ethic
- Excellent written and communication skills
- Ability to multitask, solve problems resourcefully, and work well under pressure
- Experience with strategic planning, business development, or stakeholder management
#LI-MC1
MassMutual is an equal employment opportunity employer. We welcome all persons to apply.If you need an accommodation to complete the application process, please contact us and share the specifics of the assistance you need.
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About The RoleAs a lead front-end design engineer, you will be a key part of the world-class team designing and developing the next generations of the Cerebras Wafer Scale Engine (WSE). This role requires deep expertise in RTL design and integration, with a strong focus on delivering high-performance, power-efficient, and scalable solutions. The role also requires close collaboration and management of external ASIC vendor. You will collaborate closely with the design verification, physical design, software and system teams to bring innovative semiconductor architectures from concept to production, addressing the unique challenges of building WSE systems. Responsibilities- Drive all aspects of chip design, including Functional Specification, Micro-architecture, RTL development, Synthesis.
- Managing external ASIC vendor through product development cycle.
- Work closely with PD team members for design closure to meet PPA goals.
- Work closely with Design verification and DFT teams for achieving the best functional and test coverage.
- Work with software and system teams to understand opportunities to deliver optimal performance and feature set for the product.
- Debug silicon-level functional, timing, and power issues during bring up.
- Master’s degree in Computer Science, Electrical Engineering, or equivalent.
- Can work in a hybrid work environment.
- 8-15 years of experience in delivering complex, high performance high quality RTL designs.
- Experience with Front End Chip integration and third-party IP integration.
- Demonstrated experience in networking, high-performance computing, machine learning or related fields.
- Proven track record of multiple silicon success.
- Experience collaborating and managing external vendors.
- Experience with designing/integrating high speed IO.
- Networking stack experience including TCP/IP, RDMA and Ethernet.
- Knowledge of PCIe, CPU interfaces and Serdes technology.
- Working knowledge of scripting tools : Python, TCL.
- Experience with FPGA development toolchain, including Place and Route, Floor planning and Timing Analysis is a plus.
The base salary range for this position is $175,000 to $275,000 annually. Actual compensation may include bonus and equity, and will be determined based on factors such as experience, skills, and qualifications.
This offer is contingent upon Cerebras successfully obtaining an export license from the U.S. Department of Commerce’s Bureau of Industry and Security authorizing the release to you of certain software source code and/or technology that is subject to the Export Administration Regulations. However, we can make no assurances with respect to the final disposition of an export license application.
Why Join CerebrasPeople who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
- Build a breakthrough AI platform beyond the constraints of the GPU.
- Publish and open source their cutting-edge AI research.
- Work on one of the fastest AI supercomputers in the world.
- Enjoy job stability with startup vitality.
- Our simple, non-corporate work culture that respects individual beliefs.
Read our blog: Five Reasons to Join Cerebras in 2026.
Apply today and become part of the forefront of groundbreaking advancements in AI!Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
About The RoleAs a Test Development Engineer on our manufacturing team you will be working with diagnostics, system design, manufacturing, and quality teams to develop test automation solutions for our products from PCBA to system level. You will also work closely with our contract manufacturing sites to fulfill a complete test automation solution for manufacturing test data, yield improvement, and traceability.
Responsibilities- Develop and design manufacturing test automation software/scripts to test Cerebras products from PCBA to system level.
- Develop and implement GUI solutions for test automation.
- Work with our contract manufacturers to develop and implement a test data reporting portal for manufacturing traceability and analysis.
- Sustain our current test software and infrastructure and help root cause and resolve any manufacturing test software issues or hardware defects.
- Design a web interface for user to modify/edit settings from mySQL database on AWS.
- Setup the various infrastructures at our manufacturing sites to support test equipment and server operation.
- Interact with contract manufacturing site for all the technical issues relating to manufacturing test.
- Work with diagnostics, system design, manufacturing and quality team to bring up test automation suites for the new products.
- Bachelors in computer science, electrical engineering, or other related field.
- 5+ years of experience in test automation, test development or related experience.
- Skilled in C/C++, Visual Studio, Python programming languages.
- Good knowledge of js, MySQL, SQL, SQL Server Reporting Service.
- Good knowledge of Pexpect, SSH, Telnet, RS-232, bash script.
- Good knowledge of Windows, Linux, Ubuntu, Centos, VNC viewer, Console server.
- Debugging skills and knowledge of debugging complex software stack.
- Experience in GUI development.
- Experience in Web development.
- Experience in API development.
The base salary range for this position is $170,000 to $210,000 annually. Actual compensation may include bonus and equity, and will be determined based on factors such as experience, skills, and qualifications.
Why Join CerebrasPeople who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
- Build a breakthrough AI platform beyond the constraints of the GPU.
- Publish and open source their cutting-edge AI research.
- Work on one of the fastest AI supercomputers in the world.
- Enjoy job stability with startup vitality.
- Our simple, non-corporate work culture that respects individual beliefs.
Read our blog: Five Reasons to Join Cerebras in 2026.
Apply today and become part of the forefront of groundbreaking advancements in AI!Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.
About Us:
Astiva Health, Inc., located in Orange, CA is a premier health plan provider specializing in Medicare and HMO services. With a focus on delivering comprehensive care tailored to the needs of our diverse community, we prioritize accessibility, affordability, and quality in all aspects of our services. Join us in our mission to transform healthcare delivery and make a meaningful difference in the lives of our members.
SUMMARY:
We are seeking a skilled and adaptable AI/ML Engineer to join our fast-moving team building impactful AI solutions in healthcare. Our work focuses on extracting and interpreting data from unstructured medical documents, improving clinical coding accuracy, streamlining administrative processes, and enhancing patient outreach.
Projects will evolve rapidly, from fine-tuning large language models (LLMs) on specialized medical PDFs, to optimizing OCR pipelines in Azure, and new challenges emerge regularly. This role suits someone who thrives in ambiguity, enjoys hands-on model development, and wants to directly influence healthcare delivery through applied AI/ML.
ESSENTIAL DUTIES AND RESPONSIBILITIES include the following:
- Design, fine-tune, and optimize large language models (LLMs) and multimodal models for healthcare-specific NLP tasks, such as information extraction, classification, and summarization from clinical documents (e.g., medical charts, patient files, scanned forms).
- Develop and improve document understanding pipelines, including fine-tuning OCR / layout-aware models (especially in cloud environments like Azure AI, Azure Foundry) to handle real-world variability in medical forms, handwriting, and scanned PDFs.
- Build and iterate on end-to-end ML solutions that transform unstructured healthcare data into structured, actionable insights
- Collaborate closely with clinicians, product managers, data annotators, and engineers to define problems, curate/annotate datasets, evaluate model performance against clinical and business metrics, and iterate quickly.
- Deploy models into production environments (cloud-based inference, batch processing, or API endpoints) with attention to latency, cost, scalability, and healthcare compliance considerations (HIPAA, data privacy).
- Stay current with advancements in LLMs, vision-language models, efficient fine-tuning techniques (LoRA/QLoRA, PEFT), RAG, multimodal AI, and domain-specific healthcare AI research.
- Contribute to a culture of rapid prototyping, rigorous evaluation, and continuous improvement in a dynamic project landscape where priorities can shift based on new opportunities or stakeholder needs.
- Other duties as assigned
REQUIRED TECHNICAL SKILLS:
- Strong proficiency in Python and ML frameworks (PyTorch, TensorFlow, or equivalent)
- Hands-on experience with NLP applied to unstructured text
- Experience working with LLMs, including:
- Prompting strategies
- Fine-tuning for classification or extraction tasks
- Model evaluation and error analysis
- Experience designing or consuming annotation pipelines and labeled datasets
- Familiarity with structured prediction problems (multi-label classification, ranking, or probabilistic inference)
- Ability to reason about and mitigate model bias, label noise, and false positives
- Strong understanding of production ML systems (versioning, monitoring, iteration)
- Experience working with sensitive or regulated data (e.g., HIPAA-covered healthcare data), including privacy-aware data handling and secure ML workflows
OTHER SKILLS and ABILITIES:
- Hands-on experience with Azure AI services, Azure Machine Learning, OpenAI on Azure, and Microsoft Foundry
- Experience with clinical NLP libraries (scispaCy, medspaCy, cTAKES)
- Familiarity with RAG architectures for grounding model decisions
- Experience with weak supervision or noisy-label learning
- Knowledge of temporal reasoning or longitudinal modeling
- Exposure to knowledge graphs or ontology-driven systems
- Familiarity with healthcare vocabularies and ontologies:
- ICD-10
- SNOMED CT
- RxNorm (or similar)
- Understanding of clinical documentation structure (HPI, Assessment & Plan, medications, etc.)
EXPERIENCE:
- Bachelor’s Degree in related field
- 2-4+ years of experience in software engineering, machine learning, or applied NLP
- Demonstrated experience taking ML systems from prototype to production
- Experience collaborating with non-technical domain experts (e.g., medical coders, clinicians)
BENEFITS:
- 401(k)
- Dental Insurance
- Health Insurance
- Life Insurance
- Vision Insurance
- Paid Time Off
- Free catered lunches
We have an immediate opening for a Solution Architect - GenAI with a leading IT service/solutions provider in Charlotte/Atlanta .
Job Title: Solution Architect - GenAI
Location: Charlotte/Atlanta (Hybrid 3 Days Onsite)
We are looking for a highly experienced and hands-on Solution Architect with over 10+ years of expertise in building scalable, enterprise-grade frameworks and foundational services to support next-generation Generative AI solutions.
Key Responsibilities:
- Lead and mentor a team of engineers working on full stack development of GenAI foundational services and framework.
- Architect and implement scalable and reusable infrastructure components to support enterprise-wide AI/ML initiatives.
- Define and drive the integration strategy for LLMs and Agentic AI models into core systems and frameworks.
- Collaborate with platform, data, and ML teams to define data pipelines and infrastructure for GenAI model training, deployment, and monitoring.
- Ensure GenAI solutions are technically feasible, aligned with enterprise architecture, and meet scalability and performance requirements.
- Actively participate in the design, development, and maintenance of high-performance applications using Java, Python, JavaScript, Scala, PySpark, Spark, and Snowflake.
- Work extensively on AWS services, including designing and deploying serverless or containerized solutions using ECS, Lambda, S3, Glue, etc.
- Develop and deploy robust microservices-based and event-driven architectures using modern frameworks.
- Integrate and work with GenAI frameworks and APIs such as OpenAI, Hugging Face, Google Vertex AI, or other enterprise-grade platforms.
- Establish coding standards, perform peer code reviews, and ensure adherence to best practices, including test automation and CI/CD controls.
- Ensure security compliance through vulnerability detection and remediation and implement OWASP best practices in the application lifecycle.
- Collaborate with cross-functional Agile teams including Product Owners, QA, and DevOps for sprint planning, delivery, and continuous improvement.
Cerebras Systems builds the world's largest AI chip, 56 times larger than GPUs. Our novel wafer-scale architecture provides the AI compute power of dozens of GPUs on a single chip, with the programming simplicity of a single device. This approach allows Cerebras to deliver industry-leading training and inference speeds and empowers machine learning users to effortlessly run large-scale ML applications, without the hassle of managing hundreds of GPUs or TPUs.
Cerebras' current customers include top model labs, global enterprises, and cutting-edge AI-native startups. OpenAI recently announced a multi-year partnership with Cerebras, to deploy 750 megawatts of scale, transforming key workloads with ultra high-speed inference.
Thanks to the groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution in the world, over 10 times faster than GPU-based hyperscale cloud inference services. This order of magnitude increase in speed is transforming the user experience of AI applications, unlocking real-time iteration and increasing intelligence via additional agentic computation.
The Role
We are seeking a highly skilled and motivated Manufacturing Bring-up Engineer to join our team. As the Manufacturing Bring-up Engineer you will support our system level bring-up process execution, implementation, and evolution in the manufacturing pipeline. This is a high visibility role that requires strong technical expertise, coordination, and collaboration to deliver our product from manufacturing to the customer.
Responsibilities
- Support the Cerebras manufacturing bring-up process execution to configure, test, and validate system performance prior to customer shipment
- Collaborate cross-functionally with Asic, SW, Diagnostics, and QA teams to further automate and streamline the workflow for optimal manufacturing efficiency
- Troubleshoot and resolve technical issues during system bring-up across Asic, SW, and QA domains
- Design and implement efficient processes to manage and track system bring-up status and progress
- Track and report on critical bring-up metrics to drive continuous improvement
- Implement further SW automation and efficiencies to effectively scale the manufacturing bring-up process in support of the manufacturing roadmap
Skills & Qualifications
- BS or MS in EE, ECE, CS or equivalent work experience
- 3+ years of industry experience in an operations environment
- Experience in hardware bring-up and the debug of complex systems
- Working knowledge and experience in Asic bringup and test processes
- Working knowledge of scripting in languages such as Python and/or Perl
- Proven experience in system bring-up and validation of complex computer systems or equivalent technologies
- Understanding of computer system architecture and hardware components
- Proficiency in scripting and automation tools for system bringup
- Excellent problem-solving and communication skills with the ability to work collaboratively in a fast-paced environment
- Very strong coordination and collaboration skills to manage a business-critical workflow directly in support of customer demand
Preferred:
- Familiarity in creating test and s/w infrastructure at large scale
- Working across global time zones
Location
Sunnyvale, California/ Bangalore, India/Toronto, Canada.
The base salary range for this position is $170,000 to $230,000 annually. Actual compensation may include bonus and equity, and will be determined based on factors such as experience, skills, and qualifications.
Why Join Cerebras
People who are serious about software make their own hardware. At Cerebras we have built a breakthrough architecture that is unlocking new opportunities for the AI industry. With dozens of model releases and rapid growth, we’ve reached an inflection point in our business. Members of our team tell us there are five main reasons they joined Cerebras:
- Build a breakthrough AI platform beyond the constraints of the GPU.
- Publish and open source their cutting-edge AI research.
- Work on one of the fastest AI supercomputers in the world.
- Enjoy job stability with startup vitality.
- Our simple, non-corporate work culture that respects individual beliefs.
Read our blog: Five Reasons to Join Cerebras in 2026.
Apply today and become part of the forefront of groundbreaking advancements in AI!Cerebras Systems is committed to creating an equal and diverse environment and is proud to be an equal opportunity employer. We celebrate different backgrounds, perspectives, and skills. We believe inclusive teams build better products and companies. We try every day to build a work environment that empowers people to do their best work through continuous learning, growth and support of those around them.
This website or its third-party tools process personal data. For more details, click here to review our CCPA disclosure notice.