AI

53 articles tagged with "AI"

Part 5: Results, Failure Modes, and the Path to Modern Image Generation Series

Part 5: Results, Failure Modes, and the Path to Modern Image Generation

What the GAN-based virtual try-on system achieved, where it failed (and why), the smartphone app design, and how diffusion models are changing everything for the next generation of META FIT.

Read →
Part 1: Solving the Daily Meal Planning Problem with Data Series

Part 1: Solving the Daily Meal Planning Problem with Data

How I tackled the universal 'what's for dinner' problem over a decade ago using classical data science — cleansing 20,000 recipes, 200,000 ingredient records, and nutritional data into a unified ML-ready dataset.

Read →
Part 2: Finding 'Same Nutrition, Different Meal' with Cosine Similarity Series

Part 2: Finding 'Same Nutrition, Different Meal' with Cosine Similarity

Using cosine similarity on nutritional vectors to find recipes that match a target meal's nutrition profile but offer completely different flavors — at both the recipe and menu level.

Read →
Part 4: Transforming 20,000 Recipes with ChatGPT Series

Part 4: Transforming 20,000 Recipes with ChatGPT

Years after building the original ML pipeline, LLMs changed everything — using the ChatGPT API to simplify elaborate recipes into weeknight-friendly meals, and reflecting on a decade-long journey from cosine similarity to LSTM to GPT.

Read →
Part 3: Predicting 'Non-Boring' Menus with LSTM Time Series Series

Part 3: Predicting 'Non-Boring' Menus with LSTM Time Series

Reframing meal planning as a text generation problem — using LSTM neural networks with temperature-controlled sampling to predict diverse, non-repetitive menus from historical meal sequences.

Read →
Part 1: The Beginning and the Big Picture -- The Day ChatGPT Beat the Fund Managers Series

Part 1: The Beginning and the Big Picture -- The Day ChatGPT Beat the Fund Managers

How an experiment where ChatGPT outperformed professional fund managers inspired me to build Senrigan, an AI stock price prediction service powered by a fine-tuned LLM.

Read →
Part 2: The Local GPU Challenge and Defeat -- Taking on an 8B Model with 6 GB of VRAM on the RTX 3060 Series

Part 2: The Local GPU Challenge and Defeat -- Taking on an 8B Model with 6 GB of VRAM on the RTX 3060

My first attempt at fine-tuning ELYZA Llama-3-JP-8B on a local RTX 3060 GPU, and the hard lessons learned about VRAM limitations when training large language models.

Read →
Part 3: LoRA and Quantization Explained -- The Techniques That Make LLMs Lighter Series

Part 3: LoRA and Quantization Explained -- The Techniques That Make LLMs Lighter

A deep dive into LoRA, quantization, QLoRA, and SFTTrainer -- the key techniques that enable fine-tuning large language models on consumer hardware, validated through a Mount Fuji proof-of-concept experiment.

Read →
Part 4: Stock Prediction on Colab -- Trial and Error with Three Models Series

Part 4: Stock Prediction on Colab -- Trial and Error with Three Models

Fine-tuning ELYZA 8B and LLM-jp 7.2B for stock price prediction on Google Colab, the accuracy challenges encountered, and why I ultimately pivoted to the OpenAI API.

Read →