Login Sign Up
🔔 FCM Loaded

GenAI Technical Architect

Tata Consultancy Services

2 - 5 years

Bengaluru

Posted: 20/03/2026

Getting a referral is 5x more effective than applying directly

Job Description

GenAI Technical Architect


Experience Level: - 10+ years overall IT experience, with 3+ years in AI/GenAI/LLM based solution architecture.


  • Should have strong technical expertise in Python, hands-on experience with at least one GenAI framework (LangGraph, LangChain, or Google AI Development Kit), and strong working knowledge of one hyperscaler platform (Google Cloud, Azure, or AWS).
  • The associate should lead solution design, integrating LLMs into enterprise workflows, mentoring team members, and driving production-grade implementation of GenAI use cases.
  • Good knowledge of MLOps or DevOps to automate model deployment, versioning, and monitoring.


Key Responsibilities:


Architecture & Design.

  • Design modular, scalable GenAI architectures leveraging LLMs, RAG, LangChain, LangGraph, Google ADK or Cortex Agents.
  • Define architecture patterns for multi-agent systems, context-aware pipelines, and hybrid reasoning flows.
  • Integrate LLMs (e.g., Llama, Gemini, GPT, Claude) into enterprise systems and custom applications.
  • Develop reusable prompt orchestration and workflow frameworks.
  • Establish standards for vector database (e.g., ChromaDB, Pinecone, FAISS, Weaviate, Vertex AI Matching Engine) usage, embeddings, and context retrieval.
  • Architect scalable and secure GenAI microservices leveraging cloud-native components.


Development & Implementation.

  • Lead Python-based development efforts for building prompt orchestration, tool agents, and data pipelines.
  • Develop and deploy APIs or microservices integrating LLMs with enterprise data sources.
  • Design and deploy LLM-based microservices with robust error handling, observability, and scalability.
  • Integrate custom models, open-weight models (e.g., Llama, Mistral), and API-based models (e.g., GPT, Claude, Gemini).
  • Lead the end-to-end RAG lifecycle ingestion, embedding, retrieval, generation, and evaluation.
  • Implement prompt optimization, context management, and model performance tuning.


Cloud Integration.

  • Architect, deploy and monitor GenAI workloads on one hyperscaler:
  • GCP (Vertex AI, Document AI, AlloyDB, BigQuery, Cloud Run)
  • Azure (OpenAI Service, Cognitive Search, Azure ML)
  • AWS (Bedrock, SageMaker, Lambda, API Gateway)
  • Manage cloud infrastructure for scaling AI models, ensuring cost efficiency and compliance.


Collaboration & Leadership.

  • Lead a small team of AI engineers and developers.
  • Partner with product and data teams to identify AI-driven business opportunities.
  • Conduct code reviews, enforce best practices, mentor development teams on AI/ML implementation best practices.
  • Review and optimize system designs for cost efficiency and latency performance.
  • Contribute to governance, model safety, and compliance frameworks.
  • Collaborate closely with product owners, data engineers, and business stakeholders to translate business needs into technical requirements.
  • Contribute to internal GenAI capability building and reusable assets for the organization.


Research & Innovation.

  • Stay updated on LLM research, agentic frameworks, and GenAI trends.
  • Protopye and evaluate multi-agent architectures, prompt optimization, and LLMOps pipelines.
  • Experiment with prompt engineering, fine-tuning, and model evaluation metrics.


Required Skills & Experience:


Core Technical Skills.

  • Python (advanced proficiency; ability to build APIs, pipelines, and modular frameworks).
  • Hands-on experience with RAG systems, vector databases (FAISS, Pinecone, Chroma, Weaviate, or Snowflake Cortex Search).
  • Hands-on with at least one GenAI framework:
  • LangChain, LangGraph, or Google ADK (AI Development Kit).
  • Solid understanding of LLMs (OpenAI, Anthropic, Meta, Mistral, Gemini, etc.) and token optimization strategies
  • Experience designing multi-agent or autonomous AI workflows.
  • Expertise with LLM integration (OpenAI API, Gemini API, Ollama, Hugging Face, etc.).
  • Experience with RAG, embeddings, and vector databases.
  • Familiarity with PEFT, LoRA, or prompt fine-tuning approaches.
  • Experience designing scalable microservices and event-driven architectures.
  • Proven experience in production deployment, load balancing, and monitoring AI workloads.
  • Knowledge of data engineering concepts pipelines, ingestion, metadata, and data APIs.
  • Familiarity with front-end integration (Streamlit, Gradio, or custom dashboards).
  • Cloud / Hyperscaler Expertise (at least one required)
  • Google Cloud Platform (GCP) Vertex AI, Document AI, BigQuery, AlloyDB, Cloud Run, IAM
  • Azure Azure OpenAI, Cognitive Search, Azure ML, Azure Functions
  • AWS Bedrock, SageMaker, Lambda, API Gateway, DynamoDB


Soft skills.

  • Strong analytical and problem-solving mindset.
  • Excellent communication and stakeholder management skills.
  • Proven ability to lead technical discussions and drive cross-functional alignment.


Other Desirable Skills.

  • Knowledge of REST APIs, JSON, and FastAPI/Flask frameworks.
  • Familiarity with data governance, PII handling, and AI ethics principles.
  • Understanding of Docker/Kubernetes, CI/CD, and Git-based version control.
  • Exposure to front-end integration with AI chat agents (React, Streamlit, Gradio, etc.) is a plus.
  • Offshore, open to all TCS ODC located areas

Services you might be interested in

Improve Your Resume Today

Boost your chances with professional resume services!

Get expert-reviewed, ATS-optimized resumes tailored for your experience level. Start your journey now.