Integrate the best models into your applications or deploy your own to your users, worldwide
Trusted by over 1000+ clients and partners
Designed to meet every need: LLM, image generation and classification
MODEL | TYPE | DESCRIPTION | QUANTIZATION |
---|---|---|---|
distilbert-base | Text processing | A smaller, faster version of BERT used for natural language tasks. | FP32, FP16 |
stable-diffusion | Text-to-image | Generates images from text descriptions using deep learning techniques. | FP32, FP16, INT8 |
stable-cascade | Text-to-image | Enhances image generation with multiple refinement steps. | FP32, FP16 |
sdxl-lightning | Text-to-image | Optimized for fast image generation from text inputs. | FP32, FP16 |
ResNet-50 | Image classification | A convolutional neural network designed for image recognition tasks. | FP32, FP16, INT8 |
Llama-Pro-8b | Text generation | A large language model designed for generating human-like text. | FP32, FP16, BF16 |
Llama-3.2-3B-Instruct | Text generation | An instruction-tuned model for generating text with specific guidelines. | FP32, FP16, BF16 |
Mistral-Nemo-Instruct-2407 | Text generation | Tailored for creating text based on given instructions. | FP32, FP16 |
Llama-3.1-8B-Instruct | Text generation | An advanced model for generating text with detailed instructions. | FP32, FP16, BF16 |
Pixtral-12B-2409 | Text-to-image | Produces high-quality images from text prompts using a large model. | FP32, FP16 |
Llama-3.2-1B-Instruct | Text generation | Focused on generating text according to user-provided instructions. | FP32, FP16, BF16 |
Mistral-7B-Instruct-v.0.3 | Text generation | Designed for generating guided text outputs with minimal latency. | FP32, FP16 |
Whisper-large-V3-turbo | Audio-to-text | Quickly transcribes audio into text with high accuracy. | FP32, FP16 |
Whisper-large-V3 | Audio-to-text | Transcribes spoken language into written text using deep learning. | FP32, FP16 |
Discover Sesterce playground to generate texts and images seamlessly
Get all the benefits of our platform from your terminal with our API:
Wide Range of GPU Instances - Access cutting-edge options like H200, H100 tensorCore, and more, all on-demand.
AI Inference On-Demand - Deploy your models with a dedicated endpoint, ensuring minimal latency and unlimited token pricing for seamless global access.
Unlimited Persistent Storage - Adaptable, scalable storage that grows with your needs.
No commitment, no hidden costs. You only pay for what you use.
Get real-time information about your instances and volume consumption.
Your bills available at any time, for both credit top-ups and past consumption.
Discover our awesome features to Build and Train your AI models at scale