Prince Singh

Founding Engineer @proPeers | Ex-SDE 1 | Ex-SDE Intern @CloudConduction | Mentor @proPeers & @topmate.io | LeetCode Knight πŸ‘‘ | GFG Institute Rank 1 πŸ₯‡ & Global Rank 98 🌍 | InterviewBit Global Rank 13 πŸ₯‡ | CodeStudio Specialist 🌞

Cracked National & International 4 Remote Job As A Fresher, ( 4x Remote SDE )

About

Founding Engineer with End-to-End Ownership of AI-driven features (Roadmap AI, CodeLLM, AskAI), leveraging OpenAI, ChromaDB, LangChain, MERN Stack, Docker, DevOps, and Azure. Architected self-evolving RAG pipelines, scalable CI/CD workflows, vectorized search infra, and low-latency APIs serving 100K+ users with <300ms response times. Solved 5000+ DSA problems (1200+ day streak), and mentored 40K+ Students in Systems Design, Full-Stack, and AI/ML, DSA and Problem-Solving and Interview.

Experience

ProPeers

Founding Engineer

July 2025 – Present Β· Delhi, India Β· Remote

  • β–ΉLed the launch of Roadmap AI, a fully personalized learning assistant powered by RAG (Retrieval-Augmented Generation), OpenAI’s text-embedding-ada-002, Chroma Vector DB, and Modal for real-time, scalable inference.
  • β–ΉArchitected a self-learning dynamic RAG pipeline: [JSON β†’ Embedding β†’ Chroma DB β†’ Query Context Retrieval β†’ Prompt Masking β†’ Model β†’ Nested JSON Output]
  • β–ΉDynamically decides whether to retrieve existing context or generate a roadmap from scratch, enabling zero-friction personalization for every user query.
  • β–ΉInjects prompt templates based on match confidence and automatically re-embeds new data into the vector store β€” making the system truly adaptive and self-updating.
  • β–ΉIntegrated MCP (Modular Content Pipeline) to process and vectorize 100+ roadmaps, enabling semantic search and structured AI roadmap generation.
  • β–ΉEngineered a Model Context Protocol (MCP) to standardize context injection for the model β€” combining retrieved chunks, user metadata, prompt masks, and query scaffolding β€” ensuring consistent and accurate outputs at sub-second latency.
  • β–ΉDeveloped token-based access with one-time/monthly/yearly tiers, including real-time token usage tracking, speed controls, and upsell modals for premium upgrades.
  • β–ΉAchieved <1s latency for AI responses at scale, improving retention and enabling smooth, conversational AskAI interactions.
  • β–ΉBuilt an AI-powered DSA Code Editor supporting Run/Submit/Save, tightly integrated with Roadmap AI and backed by gpt-3.5-turbo, o3-mini, and o1 models for contextual code assistance.
  • β–ΉEnhanced AskAI with contextual node + discussion integration, improving answer relevance and surfacing smarter suggestions.
  • β–ΉResulted in 3x higher roadmap completions, reduced user drop-offs, and transformed the platform into a self-evolving AI-first learning ecosystem.

SDE - 1

July 2024 – July 2025 Β· Delhi, India Β· Remote

  • β–ΉBuilt and scaled the flagship "Roadmaps" feature, delivering 100+ curated learning paths across DSA, Development, and System Design β€” used by 100K+ users. Improved personalization and relevance, while reducing API response time from 2.1s to < 300ms, resulting in a 7x faster experience and 40% higher user engagement.
  • β–ΉWorked on complex APIs to reduce processing time and improved tab switching experience for smoother navigation
  • β–ΉDeveloped and integrated the "AskAI + Discussion Forum", an intelligent peer-programming assistant where users can interact with AI to solve DSA/Dev doubts and collaborate with others β€” enabling on-demand doubt resolution and community learning.
  • β–ΉEngineered a Session Recording Bot using Python, Selenium, and headless Azure VMs with deep link automation β€” automating session joining and recording, cutting down 100% of manual effort and improving reliability.
  • β–ΉOptimized 150+ APIs by implementing advanced caching layers, async processing, and API pipelines, reducing backend latency by up to 70% and improving system throughput.
  • β–ΉReduced core web vitals TBT, LCP, and FCP from 4.4s to 990ms through advanced frontend optimizations (SSR, dynamic imports, lazy-loading APIs), significantly boosting UX for 15K+ monthly active users.
  • β–ΉLed the end-to-end performance overhaul of the platform, focusing on smoother tab-switching experiences, minimal downtime, and blazing-fast navigation across the app.
  • β–ΉMigrated MongoDB from Atlas to self-hosted replica sets, wrote automated backup & recovery scripts, set up VMs, and integrated cron-based backups to Azure Blob, ensuring data durability and cost-efficiency.
  • β–ΉSet up real-time monitoring and alerting with Prometheus and Grafana, ensuring system health, proactive issue resolution, and enhanced DevOps visibility.
  • β–ΉDeployed scalable CI/CD pipelines using Azure, GitLab, and Vercel, ensuring zero-downtime deployments and faster iteration cycles across teams.
  • β–ΉHandled end-to-end production deployment and scaling for a system serving 15K+ users, maintaining high availability, fault tolerance, and robust performance at scale.

Cloud Conduction

Junior Software Engineer

Jan 2024 – June 2024 Β· USA, Β· Remote

  • β–ΉBuilt an AI-powered chat application from the ground up using React and .NET, improving frontend efficiency by 60% and backend performance by 30%, delivering a highly responsive user experience.
  • β–ΉIntegrated and optimized AI model responses, reducing latency from 1.86s to 1.2s (35% faster) through strategic API design, caching, and performance tuning.
  • β–ΉDesigned scalable cloud architecture on Microsoft Azure for AI workloads, improving system throughput by 10% while significantly reducing infrastructure costs via autoscaling and resource optimization.
  • β–ΉDeveloped modern, responsive UI components in React that improved user engagement metrics by 25%, including better retention and interaction rates.
  • β–ΉImplemented secure, scalable API gateways in .NET Core, capable of handling 500+ concurrent requests with 99.9% uptime, supporting production-level reliability.
  • β–ΉLed the implementation of new features using the MERN stack, cutting down development time by 40%, and accelerating product iteration cycles.
  • β–ΉEstablished CI/CD pipelines (Azure DevOps & GitHub Actions), reducing deployment failures by 75% and enabling faster, automated releases.
  • β–ΉConducted in-depth code reviews and optimization, reducing technical debt by 30%, standardizing best practices across teams, and improving maintainability.
  • β–ΉOwned and managed the complete project lifecycle, from initial system design and dev planning to production deployment, server setup, and post-launch support.

Problem Solving & DSA

LeetCode

1879+ (Top 5% Worldwide)

Knight1200DaysOfCode+Annual Awards 2022/2023

1200+ problems solved | 3.5⭐ | Knight Badge

GeeksForGeeks

Global Rank Global Rank 93

1300+ problems | Institute Rank 1

InterviewBit

Global Rank Global Rank 13

560+ problems | 6⭐ Problem Solving

Other Platforms

CodeStudio

2000+ (Top 0.5%)

HackerRank

300+ problems

HackerEarth

5⭐ Python/Java

work@Tech

Global Rank 999

Key Highlights

  • β–Ή5000+ problems across 10+ platforms
  • β–Ή1200+ day unbroken streak
  • β–ΉTop 5% worldwide on multiple platforms

Technical Skills

Programming Languages

PythonJavaScriptTypeScriptC++JavaSQLBash

Frontend Development

Next.jsReactReduxTailwindCSSBootstrap

Backend Development

Node.jsExpressFastAPIDjango.NET

Cloud & DevOps

AzureAWSDockerCI/CDGrafanaPrometheus

Databases

MongoDBMySQLRedisFirebaseChromaDB

AI/ML

PyTorchLLMsRAGOpenAI APIscikit-learnVector DatabasesLangChainPrompt Engineering

Tools

GitVS CodePostmanFigmaSelenium

Education

Sage University Indore

B.Tech in Computer Science

2020 – 2024 Β· MP, India

CGPA: 8.5/10

Major Contributions ( Work Experience)

  • πŸ—ΊοΈ Created a modular content pipeline starting from JSON β†’ Embedding β†’ ChromaDB using text-embedding-ada-002.
  • πŸ”„ Designed a RAG-powered query engine that semantically retrieves roadmap context and injects it into hidden prompt templates.
  • 🧠 Applied MCP to combine RAG chunks, user metadata, and fallback strategies into OpenAI-compatible prompts.
  • πŸ“Œ Supported fallback roadmap generation without vector results to ensure resilience across sparse queries.
  • πŸ“ˆ Enabled adaptive roadmap generation for 100K+ learners β€” increasing roadmap completion rates by 3x.
  • πŸ“€ Embedded every generated roadmap back into the pipeline for future recall, creating a self-learning loop.
Low Level and High Level Architecture
  • πŸ” Developed a full-stack AI code validation engine using Retrieval-Augmented Generation (RAG), Model Context Protocol (MCP), and prompt masking for contextual understanding.
  • πŸ“¦ Fetched reference solutions via ChromaDB + text-embedding-ada-002 to provide real-time semantic support from a base of 1000+ curated DSA problems.
  • 🧠 Structured MCP Message Arrays using Problem Description (MongoDB), Reference Solution (ChromaDB), and User Code to enable precise OpenAI LLM inference.
  • βš™οΈ Enabled two-layered validation: 1) Syntax via AST Parsing, 2) Logical via chatWithO3Mini using o3-mini model.
  • 🧾 Deployed a verdict engine returning { syntax_valid, test_case_passed, approach, feedback } via structured JSON.
  • πŸš€ Achieved sub-<300ms latency through Chroma query optimization, prompt caching, and backend tuning.
  • πŸ“Š Scaled to 10K+ daily submissions using Azure auto-scaling, load-balancing, and self-healing infrastructure.
  • πŸ”„ Built a feedback loop that fine-tunes verdict accuracy monthly by analyzing user corrections.
Low Level and High Level Architecture
  • πŸ’¬ Built a dynamic conversational assistant to resolve developer doubts contextually via community threads and AI insight.
  • πŸ“š Implemented threaded conversations, follow-up suggestions, and user-personalized interaction trees.
  • πŸ“Œ Used MCP (Model Context Protocol) prompts to blend user question, system role, and learning history into single message arrays.
  • βš™οΈ Integrated token-based usage control with limit enforcement and tracking using MongoDB.
  • 🚫 Designed to run without RAG β€” answers are LLM-native and constructed through structured prompt layering alone.
  • ⚑ Delivered 3x engagement and 2x resolution speed through clean formatting (code + explanation), model-switching (O3Mini/O1), and chat memory.
Low Level and High Level Architecture
  • πŸ’» Engineered an AI-integrated code editor using Monaco, seamlessly tied into CodeLLM and AskAI pipelines.
  • βš™οΈ Supported live verdicts, multi-language (C++, Java, Python) switching, and dynamic prompts based on user activity.
  • 🧠 Embedded AI-based feedback inline within the editor via backend event sync and code stream capture.
  • πŸš€ Delivered interactive IDE-like experience with <40ms event lag, boosting engagement and retention by 40%.
  • πŸ“Š Refactored and optimized over 150 core APIs (Editor, Roadmap, AskAI, Profile) for high-throughput performance.
  • πŸš€ Reduced average response latency from 2.2s β†’ 300ms through async queues, parallel batches, and Redis caching.
  • 🧰 Introduced pagination layers, ElasticSearch indexing, and horizontal load balancing to maintain SLA under scale.
  • πŸ“ˆ Achieved 70% backend performance boost and improved Core Web Vitals (TTFB, LCP, FCP) across all pages.
  • πŸ§ͺ Load tested to 10K RPM β€” 99.95% uptime sustained with zero cold-starts using warmed cloud functions.

Β© 2025 Prince Singh. All rights reserved.

Updated at July 2025