The Reality of 'Continual Learning' ― Can AI Truly Evolve on Its Own?
As many AI services claim 'continual learning' capabilities, here's how to distinguish genuine self-learning from glorified note-taking.
11 articles tagged with "LLM"
As many AI services claim 'continual learning' capabilities, here's how to distinguish genuine self-learning from glorified note-taking.
How an experiment where ChatGPT outperformed professional fund managers inspired me to build Senrigan, an AI stock price prediction service powered by a fine-tuned LLM.
My first attempt at fine-tuning ELYZA Llama-3-JP-8B on a local RTX 3060 GPU, and the hard lessons learned about VRAM limitations when training large language models.
A deep dive into LoRA, quantization, QLoRA, and SFTTrainer -- the key techniques that enable fine-tuning large language models on consumer hardware, validated through a Mount Fuji proof-of-concept experiment.
Fine-tuning ELYZA 8B and LLM-jp 7.2B for stock price prediction on Google Colab, the accuracy challenges encountered, and why I ultimately pivoted to the OpenAI API.
A detailed look at how I designed training data for stock price prediction, integrating company information, news, stock prices, financials, and macroeconomic indicators into a structured JSON format for LLM fine-tuning.
How choosing DeepSeek as a translation provider based on cost alone led to Chinese text leakage, latency issues, and data exposure -- and why I unified everything under ChatGPT.
How pivoting to OpenAI API fine-tuning on gpt-4o-mini achieved stable JSON output and sufficient accuracy in just 8 minutes, after months of struggle with open-source models.
Local LLM inference on RTX 3060: Ollama setup and VRAM crash, LM Studio 0.4.0 headless CLI, lock mechanisms for parallel requests, mobile LLM feasibility research, and a guide for adapting to any language pair.