About the team
The Inference Ecosystem Engineering team’s mission is to show—not tell—the power of the Cerebras Inference API. We build open-source integrations, reference architectures, and polished demo apps that developers can clone, run, and extend in minutes. From LangChain agents to partner plug-ins and end-to-end “weekend projects,” our code is often the first (and most lasting) impression customers have of Cerebras.
Responsibilities:
- Design, develop, and maintain open-source libraries, SDKs, and sample repos that make Cerebras the easiest-to-adopt inference platform.
- Create production-quality demo applications that highlight low latency, high gen speed, and cost advantages.
- Build and own CI/CD pipelines, tests, and release automation for all public repos.
- Collaborate with partner engineering teams to embed Cerebras inference into their products and publish joint reference architectures.
- Collect developer feedback, identify usability gaps, and influence the Cerebras API roadmap.
- Contribute to engineering blogs, tutorials, and conference talks to grow community awareness and adoption.
Skills & Qualifications:
- Bachelor’s or Master's degree in computer science or related field, or equivalent practical experience.
- 4+ years professional software engineering experience (or equivalent open-source track record).
- Solid understanding of GenAI applications and design patterns such as RAG.
- Proficiency in Python and/or TypeScript/JavaScript
- Hands-on with at least one modern LLM framework (LangChain, LlamaIndex, CrewAI, AutoGen, etc.).
- Multiple non-trivial open-source contributions, preferably to GenAI projects
- Ability to move quickly from whiteboard idea to working prototype; bias toward shipping polished developer experiences.
- Strong communication skills—you enjoy writing clear docs and telling a compelling technical story.