Skip to main content

Mukesh Arambakam

Senior Machine Learning Engineer and Software Developer

with with 7+ years of Software Development experience, including 5 years specializing in Natural Language Processing, Neural Machine Translation, and Generative AI, backed by a Master's in Data Science.

Get In Touch Navigate to contact information and social media links View CV View detailed curriculum vitae and professional experience

Core Technologies

Python Machine Learning NLP Generative AI FastMCP PyTorch Prompt Engineering HuggingFace LangChain LangGraph

Curriculum Vitae

Download CV Link to Download CV

Professional Summary

Senior Machine Learning Engineer with with 7+ years of Software Development experience, including 5 years specializing in Natural Language Processing, Neural Machine Translation, and Generative AI, backed by a Master's in Data Science. Skilled in designing and delivering end-to-end ML solutions, including Data Preprocessing, Model Training, Evaluation and Optimization. Experience in fine-tuning LLMs with PEFT methods, building Agentic AI systems, and developing scalable, production-ready ML applications. Strong background in Machine Learning, Artificial Intelligence, Statistical Modeling, and Data Analytics.

Professional Experience

Senior Machine Learning Developer

Oracle Corporation • Dublin, Ireland

Nov. 2020 - Present
  • Launched an agentic AI system with LangGraph, supporting 10k+ LLM calls to Translation & LangDetect MCP servers.
  • Implemented 2 RAG systems with vector databases for domain-specific data extraction and translation enhancement.
  • Engineered 5+ prompt engineering solutions for back-translation, domain classification, post-editing, and data generation with sophisticated tool calling.
  • Used prompt chaining techniques for multi-step LLM workflows including translation evaluation, iterative improvement, pivot-translation and post-editing.
  • Developed a dataset analysis framework from scratch for MT training data statistics, working across 40+ language pairs.
  • Built a monolingual data generator for low-resource languages increasing synthetic data coverage by 30-40%.
  • Designed a cross-lingual vocabulary gap detection tool leveraging n-gram and embeddings, improving coverage by 20%.
  • Enhanced data-cleaning by reducing noise by 10% using alignment models, length ratios, and toxicity checks.
  • Used in-context learning methods: zero-shot and few-shot learning for domain classification & data categorization tasks.
  • Updated the Training pipeline to generate word alignment, length ratio models for data cleaning and evaluations.
  • Developed a Monolingual Engine Training pipeline and de-prioritized the tagged monolingual data during training.
  • Fine-tuned translation models using PEFT techniques for efficient adaptation to domain-specific tasks.
  • Applied quantization strategies (QLoRA) for LLM deployment, cutting resource usage by 40% in production.
  • Re-engineered translation evaluation repository with modern architectures, boosting performance by 25%.
  • Designed custom translation quality rubrics with linguists and applied LLM-as-a-Judge via direct assessment.
  • Integrated G-Eval metric and LLM-as-a-Judge framework for comprehensive LLM-based translation assessment.
  • Created visualization for evaluation scores using custom-built tools.
  • Expanded evaluation framework with 5 new metrics: Comet, XComet, BERTScore, CometQE, and length-ratio checks.
  • Assessed LLM translation consistency with Translate-then-Evaluate and LLM-as-a-Judge, raising accuracy by 20%.
  • Expanded language detection coverage from 45 to 192 languages.
  • Introduced a 3-stage hierarchical language detection architecture to differentiate sub-cultural languages.
  • Improved language detection accuracy by 90% for very short strings (<12 characters).
  • Refactored the codebase using robust system design patterns and FastAPI to significantly boost performance by 3x times.
  • Packaged and published ONNX models with architecture-aware configurations for cross-compatibility.
  • Used architecture-specific optimizations and model quantization to reduce memory footprint by 50%.

Applications Engineer

Oracle Corporation • Bangalore, India

Jun. 2019 - Aug. 2019
  • Coordinated with multiple global teams, business analysts, and stakeholders to develop a cross-platform application.
  • Lead developer in overhauling the project security flow of the Product and introduced 2 new features.
  • Implemented a micro-service which communicates between tenants in a Multi-Tenant Architecture.
  • Mentored 10 junior team members, and provided insights about the product's internal working and underlying tools.

Applications Developer

Oracle Corporation • Bangalore, India

Jul. 2017 - May. 2019
  • Certificate of Appreciation: For outstanding contribution to the Product's Development.
  • Assumed a pivotal role in the development of 3 core features of the product along with my team members.
  • Optimized the performance of an internal module's filter framework with the use of dynamic SQL queries by 30%.
  • Improved the performance of search framework by 20%, using query optimization techniques.

Education

Master of Science in Data Science

Trinity College of Dublin • Dublin, Ireland

2019 - 2020

Specialization in Data Science, Machine Learning and Artificial Intelligence

B.Tech in Computer Science and Engineering

Amrita School of Engineering • Coimbatore, India

2019 - 2020

Specialization in Data Structures, Algorithms, and System Design

Technical Skills

Programming Languages

Python Java R JavaScript C++ SQL PL/SQL

Machine Learning Libraries

HuggingFace PyTorch NLTK spaCy transformers OpenNMT scikit-learn TensorFlow fasttext

Other Python Libraries

Numpy SciPy Pandas FastAPI matplotlib elasticsearch fastalign

Generative AI & LLM Technologies

LangChain LangGraph FastMCP Agentic AI Tool Calling Prompt Engineering RAG LLM-as-a-Judge

LLM Fine-Tuning & Optimization

PEFT LoRA (Low-Rank Adaptation) QLoRA (Quantized LoRA) In-Context Learning ONNX

Infrastructure & DevOps

PEFT LoRA (Low-Rank Adaptation) QLoRA (Quantized LoRA) In-Context Learning ONNX

Projects

Home Lab Infrastructure

Sophisticated home lab using Proxmox and Docker LXCs, hosted over 10 applications including Home Assistant, Nextcloud, Immich, and Jellyfin. The environment features SSO authentication, custom DNS and DHCP for network management, and secure remote access via Tailscale VPN and Funnels.

Proxmox Docker OpenSource Deployements Networking

Image and Audio Captcha Solver

Parallelised captcha generator and solver with 93% accuracy built using TensorFlow’s CNN model.

Python Tensorflow CNN TTS OpenCV

Connect 4 - AI

A multi-agent implementation of the game Connect-4 using MCTS, Minimax and Expectimax algorithms.

Kubernetes Docker Prometheus Grafana GitLab CI Istio Ansible

Technical Blog

12 min read

Getting Started with Neural Machine Translation: A Complete Guide

Comprehensive guide to implementing neural machine translation systems from scratch using modern deep learning frameworks and best practices.

Machine Learning Read More
10 min read

Building Production-Ready ML APIs with FastAPI and Docker

Best practices for creating robust, scalable machine learning APIs using FastAPI, Docker containerization, and cloud deployment strategies.

Software Engineering Read More
15 min read

Efficient Fine-tuning of Large Language Models with LoRA

Deep dive into parameter-efficient fine-tuning techniques for large language models using LoRA and QLoRA methodologies.

Deep Learning Read More

Get In Touch

Email

amukesh.mk@gmail.com

Phone

+353 89 4960 450

Location

Dublin, Ireland

Availability

Available for consultation