Main responsibilities
- Design and develop agentic AI systems including autonomous agents, tool-using agents, multi-agent orchestration, and workflow state machines
- Design, build, and maintain AI-powered automation workflows using low-code/no-code platforms such as n8n and Make (Integromat) to orchestrate business processes, data pipelines, and cross-application integrations
- Build LLM-driven agents capable of reasoning, planning, retrieving knowledge, and executing tasks across enterprise systems
- Integrate agents with internal APIs, CRM/ERP platforms, Jira, Confluence, Slack, email, databases, payment systems, and other business tools using function/tool calling, MCP (Model Context Protocol), and A2A patterns
- Develop end-to-end AI automations that combine LLM capabilities with n8n/Make workflows to automate repetitive tasks such as document processing, lead enrichment, customer support triage, reporting, and data synchronisation across systems
- Connect AI agents to automation platforms via webhooks, API triggers, and custom nodes; manage scheduling, error handling, and conditional branching within automation workflows
- Implement tool-calling schemas, input validation, error handling, retries, rate limits, and fallback logic to ensure reliable agent execution
- Design and maintain RAG pipelines using vector databases, embedding models, reranking, and chunking strategies to ground agent outputs in enterprise knowledge
- Build safety guardrails including content filters, policy constraints, tool access controls, and human-in-the-loop approval flows for high-risk actions
- Create evaluation pipelines to measure agent reliability, task success rate, accuracy, and failure-mode behaviour using tools such as LangSmith, OpenAI Evals, or custom telemetry systems
- Implement observability and tracing of reasoning steps, tool calls, latency, cost, and error rates to support debugging and continuous improvement
- Deploy and operate agent services using Docker, Kubernetes, Terraform, and CI/CD pipelines in cloud environments (AWS, Azure, or GCP)
- Monitor agent behaviour in production, diagnose anomalies, and continuously refine agent policies and performance
- Evaluate emerging agentic AI models, frameworks, and toolkits; prototype and bench mark new approaches for scalability, robustness, and safety
- Prepare technical documentation including architecture diagrams, capability descriptions, limitations, and operational guidelines
- Communicate complex AI concepts to non-technical stakeholders and collaborate across cross-functional teams to align solutions with business needs