Applied AI engineer who ships production LLM/ML systems for real users.
Production AI Deployment | Foundation Model Integration | AI Education
I design and validate AI systems that solve real problemsโfrom model evaluation frameworks at Handshake to teaching AI fundamentals at Columbia to deploying generative models in production.
-
AI Instructor @ Columbia University (Fall 2025 - Spring 2026)
Teaching Python, AI tooling, and production workflows to justice-impacted engineers via Justice Through Code program -
AI Model Validation @ Handshake (Aug 2025 - Present)
Designed evaluation frameworks for multi-modal AI models, improving accuracy 30% through systematic prompt engineering and QA protocols
TuneStory โ Shipped 2025
Applied AI music product integrating Meta MusicGen via Modal cloud infrastructure.
Stack: MusicGen, Modal, Supabase, Gemini, TypeScript
Focus: Controllable AI generation with preserved creative intent
Key decisions:
- Chose Modal over ad-hoc GPU hosting for reproducibility + scalability
- Structured generation as modular pipelines for creative iteration
- Designed system to support future education use cases
- Spec_Tracer - AI-powered UI debugging tool with precision context capture
- jarvis_voice_agent - Multimodal voice control system (AssemblyAI + ElevenLabs)
- audio_transcriber - Speech-to-text pipeline with timestamps
AI/ML: TensorFlow, HuggingFace, Claude/LLM APIs, MusicGen, Prompt Engineering
Backend: Python, TypeScript/React, Node.js, Modal, Supabase
Data: Pandas, NumPy, SQL, data validation frameworks
Tools: Git, Docker, Notion, Figma
PhD candidate in Interactive Arts & Technology @ Simon Fraser University
Research focus: Multimodal AI systems and human-AI interaction
Teaching: AI literacy, Python fundamentals, production ML workflows
- LinkedIn: in/tylar-campbell
- Email: [email protected]
- Portfolio: tylarcampbell.com
๐ก Currently seeking: AI Product Manager or Applied AI Engineering roles where I can embed with teams to solve customer problems and own problem-to-deployment cycles.


