Documentation
  • Quick Start
  • Models
    • Block Entropy Models
    • Block Entropy Image Generation
    • Avatars
    • Smart Routing
    • Routing through Openrouter
    • Block Entropy Embeddings
  • Generation
    • Chat Completions
    • Structured Generation
    • Image generation
      • Usage
      • API Reference
    • Text-to-speech
  • Guides
    • Open Web UI
    • Silly Tavern
      • Connecting to Silly Tavern
      • Story Image Generation
    • Danswers
  • Depreciations
Powered by GitBook
On this page
  1. Models

Block Entropy Models

Block Entropy models are open source LLMs that we are currently hosting.

Block Entropy Base Models

Model: be-405b-instruct-llama3.1

This is a 4.5bpw quantization of the largest model, Llama 3.1 405B. The max context window length at the moment is 16k. This is in beta, experimental stage.

Model: be-72b-chat-qwen2

This is a 8.0bpw quantization of the Qwen2 72B parameter model. It has 16K context, 8.0bpw

Model: be-70b-instruct-llama3.1

This is a 6.0bpw of the Llama 3.1 70B model, supporting 128K context.

Model: be-70b-instruct-llama3

This is a 6.0bpw quantization of the Llama 3.0 70B parameter model at 8K context.

Block Entropy Community LoRAs or Merge Models

Model: be-120b-tessxl

This is a 4.5bpw quantization of the Tess XL model, a frankenmerge of Llama and a fine tune of goliath. It has a 12K context via NTK Rope.

Model: be-120b-goliath

This is a low rank adapter of the Goliath 120B model. It runs off the base TessXL model as a rank 64 adapter. 12K context via NTK Rope.

Model: be-120b-miqu

This is a dynamic layering of the Miqu 70B parameter model to enhance the model to 120B. It contains 8K context, at 5.0bpw

Model: be-120b-miquliz

This is a low rank adapter of the Miquliz model on top of the 120B Miqu model. While this has a higher rank (256 LoRA), this an approximation of the 120B Miquliz at 8K context, 5.0bpw.

Model: be-72b-magnumv1-qwen2

This is a 8.0bpw quantization of the Qwen2 72B parameter model. It has 8K context, 8.0bpw

Model: be-70b-cat-llama3

This is a 64 rank approximation of the Cat-llama3-instruct model at 6.0bpw.

Model: be-70b-openbiollm-llama3

This is a 128 rank approximation of the OpenBioLLM model at 8k context, 6.0bpw.

Model: be-70b-lumimaid-llama3

This is a 64 rank approximation of the Lumimaid Llama 3 model at 6.0bpw.

Model: be-70b-smaug-llama3-abliterated

This is a 64 rank approximation of the Smaug Abliterated Llama 3 model at 8K context, 6.0bpw.

PreviousModelsNextBlock Entropy Image Generation

Last updated 9 months ago