Software Engineer - AI-Native Platform

Autonomize

Autonomize

Software Engineering, Data Science

Bengaluru, Karnataka, India

Posted on Apr 16, 2026


Autonomize AI is building the intelligence layer for healthcare. Our Genesis platform replaces brittle, manual knowledge workflows with AI agents that reason, retrieve, and act - reducing administrative burden so clinicians can focus on patients.
We're looking for engineers who don't just integrate AI into software - they think in agents, design for inference, and treat LLMs as first-class infrastructure.

What This Role Is

You'll architect and ship across the full Genesis stack: agentic pipelines, backend APIs, data infrastructure, and clinical-facing UI. You'll work directly with founders and customers. You'll own things end-to-end.
This is not a role where you bolt AI onto existing CRUD. You'll be making foundational decisions about how intelligent systems are designed, evaluated, and operated at scale in a regulated industry.

You'll Thrive Here If

AI-native engineering is your default mode
You've built production systems where LLMs are doing real work - not demos, not PoCs
You've designed and shipped RAG pipelines, multi-agent workflows, or tool-using agents in production
You understand prompt engineering as an engineering discipline: versioning, evaluation, regression testing
You've instrumented AI systems for observability - latency, token usage, hallucination rate, drift
You can reason about model tradeoffs (context length, cost, latency, accuracy) and make architectural calls accordingly
You've worked with LLM SDKs (OpenAI, Anthropic, Bedrock, etc.) and agentic orchestration frameworks (LangChain, LlamaIndex, CrewAI, or similar)
You build robust backend systems
4+ years building production web applications from scratch
Deep Python proficiency; comfortable with FastAPI, Django, or Flask in production
Experience designing APIs that serve both humans and AI agents (tool schemas, structured outputs, streaming)
Async-first thinking: asyncio, task queues, event-driven architectures
Kafka, Redis, or ActiveMQ for real-time data movement
Postgres, Elasticsearch, MongoDB, or graph databases (Neo4j, TigerGraph) in production
You operate at cloud scale
Docker and Kubernetes in production - this is a hard requirement
At least one public cloud (AWS, Azure, GCP) with real operational experience
Microservices and cloud-native design patterns
You've been on-call. You know what a bad deploy feels like at 2am.
You can ship a frontend when the product demands it
React, TypeScript, or modern JS frameworks
Enough frontend fluency to build clinical interfaces without a dedicated frontend handoff

Bonus: You've Done This Before

Led a small engineering team - mentored, reviewed, unblocked
CKAD or equivalent Kubernetes certification
ML/DL model deployment experience (PyTorch, scikit-learn)
Built evaluation harnesses or used MLflow, LangSmith, or similar for AI observability
Healthcare domain experience (FHIR, HL7, clinical workflows)

What We Value Above Credentials

Bias toward action - you ship, then iterate. You learn by doing, not by planning.
Owner mentality - you don't wait for permission. You identify what needs to exist and build it.
Intellectual honesty - you'd rather say "I don't know, let me find out". When unsure, seek the right information from your peers or leaders.
Async-first communication - you write clearly, document decisions, and work well across time zones.

What You'll Get

Ground floor equity in a VC-backed healthcare AI company growing fast
Full-stack ownership - no handoffs, no silos, no "that's not my team"
Direct access to founders and customers - your technical decisions will be seen and felt
Professional development budget - conferences, courses, certifications, books
Flexible, remote-friendly culture built around output, not hours
If you've been waiting for a role where AI isn't a feature - it's the foundation - this is it.