Instrumentl logo

Software Engineer, AI/ML GenAI

Instrumentl

Job Description

About Instrumentl

Hello, we’re Instrumentl. We’re a mission-driven startup helping the nonprofit sector to drive impact, and we’re well on our way to becoming the #1 most-loved grant discovery and management tool.

About us: Instrumentl is a hyper growth YC-backed startup with over 4,000 nonprofit clients, from local homeless shelters to larger organizations like the San Diego Zoo and the University of Alaska. We are building the future of fundraising automation, helping nonprofits to discover, track, and manage grants efficiently through our SaaS platform. Our charts are dramatically up-and-to-the-right 📈 — we’re cash flow positive and doubling year-over-year, with customers who love us (NPS is 65+ and Ellis PMF survey is 60+). Join us on this rocket ship to Mars!

About the Role

As a Software Engineer, AI/ML Gen AI at Instrumentl, you’ll own the full lifecycle of AI features—from rapid prototyping to production deployment and ongoing evaluation.

You will build agentic LLM systems that can plan and use tools, implement RAG pipelines over our domain data, manage and evolve embeddings, and stand up evaluation/observability so our AI is grounded, safe, and cost‑effective.

You’ll embed with one of the product pods in a hands-on role, collaborating closely with Product and Design, while partnering with DTI on platform‑level AI capabilities.

The Instrumentl team is fully distributed (though if you’d like to work from our Oakland office, we would love to see you there). For this position, we are looking for someone who has overlap with Pacific Time Zone working hours.

What you will do

  • Design agentic systems & ship AI to production: Build resilient, observable services, while optimizing cost and latency budgets. Build tool‑using LLM “agents” (task planning, function/tool calling, multi‑step workflows, guardrails) for tasks like grant discovery, application drafting, document parsing and many more.
  • Own RAG end‑to‑end: Ingest and normalize content, choose chunking/embedding strategies, implement hybrid retrieval, re‑ranking, citations, and grounding. Continuously improve recall/precision.
  • Manage embeddings at scale: Select, evaluate, and migrate embedding models; maintain vector stores (e.g., pgvector/Qdrant/Pinecone etc.); monitor drift and rebuild strategies.
  • Collaborate cross‑functionally while raising engineering standards: Work side by side with Product, Design on scoping, UX, and measurement; run experiments (A/B, canaries), interpret results, and iterate. Write clear, maintainable code, add tests and docs, and contribute to reliability practices (alerts, dashboards, incident response).

What we're looking for

  • Software engineering background: 5+ years of professional software engineering experience (as an IC), including 2+ years working with modern LLMs.
  • Proven production impact: You’ve taken LLM/RAG systems from prototype to production, owned reliability/observability, and iterated post‑launch based on evals and user feedback.
  • LLM agentic systems: Experience building tool/function‑calling workflows, planning/execution loops, and safe tool integrations (e.g., with Lang Chain/Lang Graph, Llama Index, Semantic Kernel, or custom orchestration).
  • RAG expertise: Strong grasp of document ingestion, chunking/windowing, embeddings, hybrid search (keyword + vector), re‑ranking, and grounded citations. Experience with re‑rankers/cross‑encoders, hybrid retrieval tuning, or search/recommendation systems.
  • Embeddings & vector stores: Hands‑on with embedding model selection/versioning and vector DBs (e.g., pgvector, Qdrant, Pinecone, Weaviate, Milvus etc.).
  • Evaluation mindset: Comfort designing eval suites (RAG/QA, extraction, summarization), using automated and human‑in‑the‑loop methods; familiarity with frameworks like Ragas/Deep Eval/Open AI Evals or equivalent.
  • Infrastructure & languages: Proficiency in Python (Fast API, Celery); Experience with GCP/AWS, Docker, CI/CD, and observability (logs/metrics/traces).
  • Data chops: Comfortable with SQL, schema design, and building/maintaining data pipelines that power retrieval and evaluation.
  • Collaborative approach: You thrive in a cross‑functional environment and can translate research ideas into shippable, user‑friendly features.
  • Results‑driven: Bias for action and ownership with an eye for speed, quality, and simplicity.

Nice to have

  • Startup Experience and comfort operating in fast, scrappy environments is a plus.
  • Familiarity with responsible AI, red‑teaming, and domain‑specific safety policies.
  • Fine‑tuning: Practical experience with SFT/LoRA or instruction‑tuning (and good intuition for when fine‑tuning vs. prompting vs. model choice is the right lever).

Compensation & Benefits

  • Salary ranges are based on market data, relative to our size, industry, and stage of growth. Salary is one part of total compensation, which also includes equity, perks, and competitive benefits.
  • For US-based candidates, our target salary band is $175,000 - $220,000/year + equity. Salary decisions will be based on multiple factors including geographic location, qualifications for the role, skills, proficiency, and experience level.
  • 100% covered health, dental, and vision insurance for employees, 50% for dependents.
  • Generous PTO policy, including parental leave.
  • 401(k).
  • Company laptop + stipend to set up your home workstation.
  • Company retreats for in-person time with your colleagues.
  • Work with awesome nonprofits around the US. We partner with incredible organizations doing meaningful work, and you get to help power their success.

Share this job:

Apply for this job

Please mention you found this job on Remote Nomad Jobs. It helps us keep sharing more great jobs!

    © Copyright Remote Nomad Jobs 2026