Seekr Technologies offers hosted large language model (LLM) inference services through their SeekrGenerate product, which pairs LLMs with Retrieval Augmentation Generation (RAG) for personalized customer experiences. Additionally, SeekrFlow simplifies the alignment process for enterprises to fine-tune AI to their desired principles, values, and industry regulations, indicating that they provide finetuning services for large language models.
Offer hosted llm inference and finetuning
Continue offers continuous fine-tuning of custom models using development data and supports deployment options including locally, in the cloud, or with SaaS, which implies they offer hosted LLM inference services. Additionally, they enable continuous fine-tuning of custom models.
Giga ML offers a platform that supports Mistral-7B and Mixtral, claiming to be the world's fastest inference engine for LLMs, which indicates they provide hosted LLM inference services. Additionally, they enable fine-tuning on smaller LLMs and offer fine-tuning for developers, meeting the criteria for providing finetuning services for large language models.
Anarchy offers a platform that includes fine-tuning public models and provides support services that facilitate fine-tuning and distillation. Additionally, it offers open-source LLM training infrastructure and tooling for developers, which implies hosted LLM inference services.
Salad Technologies offers hosted large language model (LLM) inference services through SaladCloud, which supports language models and provides a fully managed container service. Additionally, SaladCloud's Dreambooth API offers finetuning services for models like Stable Diffusion.
Together AI offers hosted large language model (LLM) inference services through their Together Inference product and provides finetuning services through their Together Fine-tuning product.