Gen Ai Engineer
Sharc Hire
2 - 5 years
Mumbai
Posted: 08/03/2026
Job Description
Engagement Details:
- Contract Duration: Min 36 months
- Work Timing: 8:00 AM 4:00 PM EST
- Start Timeline: Within 2 weeks
Position Overview
We are seeking experienced Data/GenAI Engineers to join our Professional Services
team on a contract basis. You will work directly on client engagements delivering
production-grade Generative AI solutions, including conversational AI assistants,
document processing automation, RAG (Retrieval-Augmented Generation) systems,
and AI-powered data analytics platforms. This role requires hands-on technical
execution, client interaction, and the ability to work independently within an agile
delivery framework.
Primary Responsibilities
GenAI Solution Development
Design and implement production-ready Generative AI applications using
Amazon Bedrock, Anthropic Claude, and other foundation models
Build and optimize RAG (Retrieval-Augmented Generation) pipelines with vector
databases (Weaviate, OpenSearch, Pinecone)
Develop AI agents and multi-agent orchestration systems using frameworks like
LangChain, LlamaIndex, or custom implementations
Create conversational AI interfaces with natural language understanding, intent
detection, and context management
Implement prompt engineering strategies, few-shot learning, and fine-tuning
approaches for domain-specific applications
AWS Cloud Architecture & Development
Build serverless architectures using AWS Lambda, API Gateway, Step Functions,
and EventBridge
Design and implement data pipelines for AI model training, inference, and
feedback loops
Develop RESTful APIs and WebSocket connections for real-time AI interactions
Configure and optimize AWS services including S3, DynamoDB, RDS, SQS,
SNS, and CloudWatch
Implement infrastructure-as-code using CloudFormation, CDK, or Terraform
Data Engineering & ML Operations
Design and build data ingestion pipelines for structured and unstructured data
sources
Implement ETL/ELT workflows for data preparation, cleaning, and transformation
Create vector embeddings and semantic search capabilities for knowledge
retrieval
Develop data validation, quality monitoring, and observability frameworks
Optimize model inference performance, latency, and cost efficiency
Client Engagement & Delivery
Participate in sprint planning, daily standups, and client review sessions
Translate business requirements into technical specifications and implementation
plans
Provide technical guidance and recommendations to clients on AI/ML best
practices
Document architecture decisions, code, and deployment procedures
Troubleshoot production issues and implement solutions quickly
Required Technical Skills (Priority Order)
Tier 1 - Critical Must-Haves
Amazon Bedrock - Hands-on experience with foundation models (Claude, Nova,
Llama or others), model invocation, streaming responses, and guardrails
Agent Frameworks & Orchestration - Production experience with LangChain,
LlamaIndex, Bedrock Agents, or custom multi-agent orchestration systems
Python - Advanced proficiency with modern Python (3.9+), including async/await,
type hints, and testing frameworks (pytest, unittest)
AWS Lambda & Serverless - Production experience building event-driven
architectures, function optimization, and cold start mitigation
Vector Databases - Practical experience with at least one: Weaviate,
OpenSearch, Pinecone, Chroma, or FAISS for semantic search
LLM Integration - Direct experience with LLM APIs (Anthropic, OpenAI, Cohere),
prompt engineering, and response parsing
API Development - RESTful API design and implementation using FastAPI,
Flask, or similar frameworks
Tier 2 - Highly Valuable
Amazon Bedrock AgentCore - Experience with AgentCore Runtime, Memory,
Gateway, and Observability for building production agent systems
AWS API Gateway - Configuration, authorization, throttling, and integration with
Lambda/backend services
DynamoDB - NoSQL data modeling, single-table design, GSI/LSI optimization,
and DynamoDB Streams
AWS Step Functions - Workflow orchestration for complex AI pipelines and
multi-step processes
Docker & Containers - Containerization, ECR, ECS/Fargate deployment for AI
workloads
Data Processing - Experience with Pandas, PySpark, AWS Glue, or similar data
transformation tools
Tier 3 - Strong Differentiators
RAG Architecture - End-to-end RAG system design including chunking
strategies, retrieval optimization, and context management
Embedding Models - Working knowledge of text embeddings (Bedrock Titan,
OpenAI, Cohere) and embedding optimization
AWS S3 & Data Lakes - S3 event notifications, lifecycle policies, and data lake
architecture patterns
CloudWatch & Observability - Logging, metrics, alarms, and distributed tracing
for AI applications
IAM & Security - AWS security best practices, least privilege access, secrets
management (Secrets Manager, Parameter Store)
CI/CD Pipelines - Experience with CodePipeline, GitHub Actions, or GitLab CI for
automated deployments
Tier 4 - Nice to Have
SageMaker - Model training, deployment, endpoints, and feature stores
OpenSearch - Full-text search, vector search, and hybrid search implementations
EventBridge - Event-driven architectures and cross-service integrations
WebSockets - Real-time bidirectional communication for streaming AI responses
AWS CDK - Infrastructure-as-code using Python or TypeScript CDK constructs
Fine-tuning & Training - Experience with model fine-tuning, PEFT methods, or
custom model training
Required Experience & Qualifications
5+ years of software engineering experience with at least 2+ years focused on
AI/ML, data engineering, or cloud-native development
2+ years of hands-on AWS experience with production deployments
1+ years of direct Generative AI experience (LLMs, embeddings, RAG, agents)
Proven track record delivering production AI applications from concept to
deployment
Strong understanding of software engineering best practices (version control,
testing, code review, documentation)
Experience working in agile/scrum environments with distributed teams
Excellent problem-solving skills and ability to work independently with minimal
supervision
Strong written and verbal communication skills for client-facing interactions
Preferred Qualifications
AWS Certifications: Solutions Architect Associate/Professional, Machine
Learning Specialty, or Developer Associate
Background in healthcare, financial services, or regulated industries with
understanding of compliance requirements (HIPAA, PCI-DSS, SOC 2)
Contributions to open-source AI/ML projects or published technical content
Experience with multi-tenant SaaS architectures and data isolation patterns
Knowledge of cost optimization strategies for AI workloads (model selection,
caching, batching)
Familiarity with frontend frameworks (React, Angular) for building AI-powered
UIs.
Project Examples You May Work On
Building conversational AI assistants for customer service automation using
Bedrock and Anthropic Claude
Implementing RAG systems for document processing, classification, and
intelligent search
Developing AI-powered data extraction and validation pipelines for healthcare
claims processing
Creating multi-agent systems for complex workflow automation and decision
support
Building integration marketplaces connecting AI capabilities to third-party
platforms
Designing voice AI solutions using Amazon Connect and Polly for customer
engagement
- Implementing AI-driven content recommendation and personalization engines.
Services you might be interested in
Improve Your Resume Today
Boost your chances with professional resume services!
Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.
