LOG IN
SIGN UP
Tech Job Finder - Find Software, Technology Sales and Product Manager Jobs.
Sign In
OR continue with e-mail and password
E-mail address
Password
Don't have an account?
Reset password
Join Tech Job Finder
OR continue with e-mail and password
E-mail address
First name
Last name
Username
Password
Confirm Password
How did you hear about us?
By signing up, you agree to our Terms & Conditions and Privacy Policy.

Technology Support II

at J.P. Morgan

Back to all Cloud & DevOps jobs
J.P. Morgan logo
Bulge Bracket Investment Banks

Technology Support II

at J.P. Morgan

JuniorNo visa sponsorshipAWS/GCP/Azure DevOps

Posted 22 days ago

No clicks

Compensation
Not specified

Currency: Not specified

City
Mumbai
Country
India

Join the Asset & Wealth Management Technology Support team to ensure operational stability, availability, and performance of production application flows. You will troubleshoot incidents, automate operational tasks with Python, implement observability (Prometheus, Grafana, CloudWatch), and define SLIs/SLOs. The role supports LLM and AI/ML deployments, deploys and monitors workloads on AWS/Kubernetes, manages batch schedulers like Autosys/Control-M, writes SQL for data tasks, and participates in on-call rotations and postmortems. Requires strong SRE/DevOps skills and a minimum of 2+ years of applied experience.

Location: Mumbai, Maharashtra, India

Join a dynamic team shaping the tech backbone of our operations, where your expertise fuels seamless system functionality and innovation.

As a Technology Support II team member in Asset & Wealth Management, you will play a vital role in ensuring the operational stability, availability, and performance of our production application flows. Your efforts in troubleshooting, maintaining, identifying, escalating, and resolving production service interruptions for all internally and externally developed systems support a seamless user experience and a culture of continuous improvement.

Job responsibilities

  • Develops automation scripts and tools in Python to streamline operations and incident response.
  • Implements and maintain observability solutions (e.g., Prometheus, Grafana, CloudWatch) to monitor system health, collect metrics, and enable proactive incident detection. Defines, measure, and report on Service Level Indicators (SLIs) and Service Level Objectives (SLOs) for critical services.
  • Supports the deployment, monitoring, and reliability of Large Language Model (LLM) applications and systems utilizing Model Context Protocol (MCP).
  • Ensures high availability and reliability of applications running on modern infrastructure such as AWS, Kubernetes, and related cloud-native platforms, and batch processing environments.
  • Deploys, monitor, and troubleshoot workloads on AWS, leveraging cloud-native services.
  • Manages, monitor, and optimize batch jobs using schedulers like Autosys and Control-M.
  • Writes and optimize SQL queries for data extraction, transformation, and reporting.
  • Participates in on-call rotations, respond to production incidents, and drive root cause analysis and postmortems. Works closely with data science, engineering, and operations teams to support AI/ML model deployment, LLM workflows, and batch processes.
  • Identifies reliability gaps and drive initiatives to improve system resilience, scalability, and efficiency.

Required qualifications, capabilities, and skills

  • Formal training or certification on software engineering concepts and 2+ years applied experience
  • Proven experience as an SRE, DevOps Engineer, or similar role supporting AI/ML, LLM, and batch processing environments.
  • Exposure to Large Language Models (LLMs) and Model Context Protocol (MCP). Proficiency in Python for automation and scripting.
  • Strong knowledge of AWS cloud services and infrastructure.
  • Experience with SQL and relational databases.
  • Hands-on experience with job schedulers (Autosys, Control-M, or similar).
  • Familiarity with observability and telemetry tools (e.g., Prometheus, Grafana, CloudWatch, Datadog).
  • Understanding of SLI/SLO concepts and their application in production environments.
  • Solid troubleshooting and incident management skills.

Preferred qualifications, capabilities, and skills

  • Experience supporting AI/ML and LLM model deployment and monitoring.
  • Exposure to containerization and orchestration (Docker, Kubernetes).
  • Knowledge of CI/CD pipelines and infrastructure as code (Terraform, CloudFormation).
  • Experience with other cloud platforms (Azure, GCP) is a plus.

     


 

Ensure optimal tech performance in a key support role, driving stability and efficiency at a leading global financial firm.

Technology Support II

at J.P. Morgan

Back to all Cloud & DevOps jobs
J.P. Morgan logo
Bulge Bracket Investment Banks

Technology Support II

at J.P. Morgan

JuniorNo visa sponsorshipAWS/GCP/Azure DevOps

Posted 22 days ago

No clicks

Compensation
Not specified

Currency: Not specified

City
Mumbai
Country
India

Join the Asset & Wealth Management Technology Support team to ensure operational stability, availability, and performance of production application flows. You will troubleshoot incidents, automate operational tasks with Python, implement observability (Prometheus, Grafana, CloudWatch), and define SLIs/SLOs. The role supports LLM and AI/ML deployments, deploys and monitors workloads on AWS/Kubernetes, manages batch schedulers like Autosys/Control-M, writes SQL for data tasks, and participates in on-call rotations and postmortems. Requires strong SRE/DevOps skills and a minimum of 2+ years of applied experience.

Location: Mumbai, Maharashtra, India

Join a dynamic team shaping the tech backbone of our operations, where your expertise fuels seamless system functionality and innovation.

As a Technology Support II team member in Asset & Wealth Management, you will play a vital role in ensuring the operational stability, availability, and performance of our production application flows. Your efforts in troubleshooting, maintaining, identifying, escalating, and resolving production service interruptions for all internally and externally developed systems support a seamless user experience and a culture of continuous improvement.

Job responsibilities

  • Develops automation scripts and tools in Python to streamline operations and incident response.
  • Implements and maintain observability solutions (e.g., Prometheus, Grafana, CloudWatch) to monitor system health, collect metrics, and enable proactive incident detection. Defines, measure, and report on Service Level Indicators (SLIs) and Service Level Objectives (SLOs) for critical services.
  • Supports the deployment, monitoring, and reliability of Large Language Model (LLM) applications and systems utilizing Model Context Protocol (MCP).
  • Ensures high availability and reliability of applications running on modern infrastructure such as AWS, Kubernetes, and related cloud-native platforms, and batch processing environments.
  • Deploys, monitor, and troubleshoot workloads on AWS, leveraging cloud-native services.
  • Manages, monitor, and optimize batch jobs using schedulers like Autosys and Control-M.
  • Writes and optimize SQL queries for data extraction, transformation, and reporting.
  • Participates in on-call rotations, respond to production incidents, and drive root cause analysis and postmortems. Works closely with data science, engineering, and operations teams to support AI/ML model deployment, LLM workflows, and batch processes.
  • Identifies reliability gaps and drive initiatives to improve system resilience, scalability, and efficiency.

Required qualifications, capabilities, and skills

  • Formal training or certification on software engineering concepts and 2+ years applied experience
  • Proven experience as an SRE, DevOps Engineer, or similar role supporting AI/ML, LLM, and batch processing environments.
  • Exposure to Large Language Models (LLMs) and Model Context Protocol (MCP). Proficiency in Python for automation and scripting.
  • Strong knowledge of AWS cloud services and infrastructure.
  • Experience with SQL and relational databases.
  • Hands-on experience with job schedulers (Autosys, Control-M, or similar).
  • Familiarity with observability and telemetry tools (e.g., Prometheus, Grafana, CloudWatch, Datadog).
  • Understanding of SLI/SLO concepts and their application in production environments.
  • Solid troubleshooting and incident management skills.

Preferred qualifications, capabilities, and skills

  • Experience supporting AI/ML and LLM model deployment and monitoring.
  • Exposure to containerization and orchestration (Docker, Kubernetes).
  • Knowledge of CI/CD pipelines and infrastructure as code (Terraform, CloudFormation).
  • Experience with other cloud platforms (Azure, GCP) is a plus.

     


 

Ensure optimal tech performance in a key support role, driving stability and efficiency at a leading global financial firm.