OtotaO_unsloth-mcp-server

OtotaO_unsloth-mcp-server

3.1

If you are the rightful owner of OtotaO_unsloth-mcp-server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.

An MCP server for Unsloth - a library that makes LLM fine-tuning 2x faster with 80% less memory.

Unsloth is a library designed to enhance the efficiency of fine-tuning large language models. It offers significant improvements in speed and memory usage, allowing for faster fine-tuning and the ability to handle larger models on consumer-grade GPUs. Unsloth supports extended context lengths and maintains model quality without performance loss. These advancements are achieved through custom CUDA kernels, optimized backpropagation, and dynamic 4-bit quantization. The library is particularly beneficial for models like Llama, Mistral, Phi, and Gemma, providing a simple API for model operations and export capabilities to various formats.

Features

  • Optimize fine-tuning for Llama, Mistral, Phi, Gemma, and other models
  • 4-bit quantization for efficient training
  • Extended context length support
  • Simple API for model loading, fine-tuning, and inference
  • Export to various formats (GGUF, Hugging Face, etc.)

Tools

  1. check_installation

    Verify if Unsloth is properly installed on your system.

  2. list_supported_models

    Get a list of all models supported by Unsloth, including Llama, Mistral, Phi, and Gemma variants.

  3. load_model

    Load a pretrained model with Unsloth optimizations for faster inference and fine-tuning.

  4. finetune_model

    Fine-tune a model with Unsloth optimizations using LoRA/QLoRA techniques.

  5. generate_text

    Generate text using a fine-tuned Unsloth model.

  6. export_model

    Export a fine-tuned Unsloth model to various formats for deployment.