Cloud Security Engineer – Cloud/AI (LLM Security)

Company: Gazelle Global
Apply for the Cloud Security Engineer – Cloud/AI (LLM Security)
Location: Norwich
Job Description:

Overview

  • Deliver cloud security engineering capability focused on securing AI, LLM, and cloud‑native workloads, with AWS as the primary environment and Azure as a secondary platform.
  • Implement secure cloud architectures and controls, ensuring AI/LLM workloads comply with organisational security standards and cloud security policies.
  • Work with architects and AI engineering teams to define secure patterns for LLM deployments, AI agents, and model pipelines across cloud environments.
  • Engineer and operationalise cloud‑native security tooling, including IaC security, secrets management, container security, and monitoring solutions.
  • Integrate security controls into CI/CD pipelines and modern development workflows, enabling secure and automated deployment of cloud and AI workloads.
  • Participate in threat modelling, risk assessment, and security design reviews for AI applications, APIs, and cloud services.
  • Support evaluation and onboarding of emerging AI security tools and cloud-native security capabilities, contributing to technology selection and capability uplift.

Qualifications

  • Strong background in cloud security engineering, ideally with deep experience on AWS; Azure exposure is highly beneficial.
  • Hands-on experience or working exposure in securing LLM/AI workloads, including model deployment, data flows, and runtime considerations.
  • Proficiency with cloud-native security tooling (CSPM, CWPP, secrets management, logging/monitoring, container security).
  • Experience securing IaC and CI/CD pipelines using tools such as Terraform, CloudFormation, GitHub Actions, GitLab, or similar.
  • Knowledge of IAM design, network security controls, encryption, secrets management, and cloud identity principles.
  • Understanding of modern cloud architectures (serverless, microservices, managed AI/ML services) and their associated security risks.
  • Ability to collaborate effectively with AI engineers, developers, and cloud teams to ensure secure implementation of AI workloads.
  • Experience securing GenAI, Agentic AI, vector databases, model APIs, or data pipelines used by LLMs.
  • Knowledge of responsible AI principles, model governance, or AI-specific threat modelling (e.g., adversarial ML, data poisoning, prompt injection).
  • Background working in regulated industries such as Financial Services or Insurance.
  • Strong stakeholder communication skills, including the ability to influence engineering teams and articulate cloud/AI security risks clearly.

#J-18808-Ljbffr…

Posted: March 30th, 2026