🌟 New service One-Click LLM Deployment. → Read more

LLM as a service

Fine-tuned SEA language models, Text-to-SQL, efficient tokenization, and cost-effective API integration.

Chat

SeaLLM-7b-v2.5

user

User

Hello

robot

Float16

Hello! How can I assist you today?

We support the models

Feature comparison

Price

per million token

Model Name

SeaLLM-7b-v3

$0.2

SeaLLM-7b-v2.5

$0.2

SQLCoder-7b-2

$0.6

Use case

Model Name

SeaLLM-7b-v3

RAG, NER, Sentiment analysis

SeaLLM-7b-v2.5

RAG, NER, Sentiment analysis

SQLCoder-7b-2

Text-to-SQL

Learn more

LLM as a service

Seamless LLM Integration

Your Quick Setup Guide

1.

Log in to Float16.cloud and navigate to LLM as a Service.

curl -X POST https://api.float16.cloud/v1/chat/completions -d 

  '{
    "model": "SeaLLM-7b-v2.5",
    "messages": [
      {
        "role": "system",
        "content": "You are a helpful assistant."
      },
      {
        "role": "user",
        "content": "สวัสดี"
      }
    ]
   }'

  -H "Content-Type: application/json" 
  -H "Authorization: Bearer <float16-api-key>"

3.

Enter your float16-api-key and model.

4.

Connect to your app via our API.

Pay per token

Cost-efficient pricing model based on actual tokens used, ensuring you only pay for what you need.

OpenAI API Compatibility

Our models are compatible with OpenAI API, allowing you to easily integrate our models into your existing workflow.

Context caching

Our models are optimized for context caching, allowing you to reuse context across multiple request for faster response times.