Experiment - Self-hosted LLMs (Llama, Mistral)
Problem:You want to run a large language model (LLM) like Llama or Mistral on your own computer or server to generate text without relying on external APIs.
Current Metrics:Model loads successfully but generates repetitive or low-quality text. Response time is slow. No fine-tuning or prompt optimization applied.
Issue:The model overfits on small prompts and produces repetitive outputs. Performance is slow due to inefficient settings. Quality and speed need improvement.