Architecting Intelligence

Rahmad Kurniawan

Lead AI Engineer | NLP & Machine Learning

Building production-grade AI systems, from tokenization to LLM-powered products at scale.

Genesis

6+ Years of
AI/ML experience

A technical lead specializing in the intersection of NLP research and robust engineering. I bridge the gap between theoretical machine learning and production-ready applications, with extensive experience in LLM orchestration, transformer optimization, and large-scale MLOps.

Architectures

Transformers, BERT, GPT-Like Model, Sequence Models

Intelligence

LLM Orchestration, RAG, Fine-tuning, NLP Pipeline Optimization

Systems

LLMOps, Scalable Inference, Vector Databases, AWS/GCP AI Infrastructure

1M+

MAU Ecosystem Scale

45%

Operational Cost Reduction

26%

Model Quality Improvement

Timeline

2022 — PRESENT

KitaLulus

Data Scientist Lead

Leading AI innovation for a high-growth career platform. Architected the AI interview SaaS, optimizing candidate matching through proprietary LLM pipelines.

  • check_circle Orchestrated LLMOps for automated interview evaluation systems.
  • check_circle Building recommendation system for job-candidate matching.
  • check_circle Scaled ML services to handle over 1M monthly active users.
2020 — 2022

Widya Wicara

Tech Lead & ML Engineer

Headed the Speech AI division, focusing on Automatic Speech Recognition (ASR) and Text-to-Speech (TTS) systems for enterprise clients.

  • check_circle Implemented ASR/TTS architectures reducing monthly operational costs by $3,500.
  • check_circle Directed technical strategy for neural voice synthesis products.
  • check_circle Mentored a team of ML engineers and Hardware Engineers.
Selected Works

System Deployments

AI Interview SaaS

arrow_outward
Problem

Inefficient manual screening for thousands of job applications.

Approach

Deployed AI Voice Interview Agent for candidate assessment.

Impact

70% reduction in initial screening time per candidate.

PyTorch HuggingFace AWS

Enterprise ASR and TTS

arrow_outward
Problem

High latency and cost in third-party speech recognition APIs.

Approach

Built end-to-end speech pipeline with self-hosted ASR and neural TTS models, pretrained and fine-tuned on localized Indonesian dialects for low-latency enterprise deployments.

Impact

$3,500/mo cost savings with 15% better accuracy.

Wav2Vec2 FastAPI Nvidia Triton

MLOps Orchestrator

arrow_outward
Problem

Fragmented model training and deployment cycles across teams.

Approach

Designed custom training pipelines with automated retraining triggers, model versioning, and canary deployments for continuous model improvement.

Impact

Deployment frequency increased by 4x with zero downtime.

Kubernetes Kubeflow Docker

Vector Search Engine

arrow_outward
Problem

Traditional keyword search failing for complex user intent.

Approach

Semantic vector embeddings using Pinecone and bi-encoder models.

Impact

26% improvement in search relevance and user satisfaction.

Pinecone Sentence-BERT Python

Core Competencies

neurology

NLP & LLM

Prompt Engineering Transformers Fine-tuning LangChain
settings_voice

Speech AI

ASR TTS Spectrogram Analysis Espnet
developer_board

MLOps

Kubernetes MLFlow CI/CD BentoML
code

Programming

Python C++ Go PyTorch

Let’s build something impactful with AI

Open for collaborations on production LLM systems, RAG architectures, and technical leadership roles.