For research, you can do things like PSO or Genetic algorithm optimizations. Those benefit more from CPU than GPU cause there is not so much matrix math going on, especially for complex fitness functions that require sequences.
You could do this with the small Llama model, where the fitness function is basically the ability generate correct code and self detect errors, and adjust the weights based on the optimization algorithm.