Hyperparameter tuning is essential for optimizing large language models (LLMs), as parameters like learning rate, batch size, weight decay, and training epochs significantly influence model performance and convergence. Traditional manual methods such as grid or random search are inefficient for compute-intensive tasks. By automating this process using CircleCI matrix jobs and integrating Weights & Biases (wandb) for tracking, users can streamline experiment tracking and reproducibility. This setup involves creating a reproducible training script that accepts command-line arguments for hyperparameters, which are then handled by CircleCI to initiate parallel jobs with different parameter combinations. The results are logged in real time to wandb, providing a central dashboard to compare experiments. This system simplifies the hyperparameter tuning process, making it scalable and automated, and can be triggered by code changes or schedules. This method greatly reduces manual effort and enhances the ability to scale experiments across different environments.