NeverSleep: Lumimaid v0.2 8B

Text input Text output
Author's Description

Lumimaid v0.2 8B is a finetune of [Llama 3.1 8B](/models/meta-llama/llama-3.1-8b-instruct) with a "HUGE step up dataset wise" compared to Lumimaid v0.1. Sloppy chats output were purged. Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).

Key Specifications
Cost
$$$
Context
32K
Parameters
8B
Released
Sep 14, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Logit Bias Structured Outputs Response Format Stop Seed Min P Top P Max Tokens Frequency Penalty Temperature Presence Penalty
Features

This model supports the following features:

Response Format Structured Outputs
Performance Summary

NeverSleep: Lumimaid v0.2 8B demonstrates exceptional performance in terms of operational efficiency, consistently ranking among the fastest models and offering highly competitive pricing across all benchmarks. This makes it a cost-effective and responsive option for various applications. However, its accuracy across most benchmark categories is notably low. In Hallucinations (Baseline), it achieved 60.0% accuracy, indicating a tendency to generate information without acknowledging uncertainty. General Knowledge and Ethics benchmarks showed particularly weak performance at 20.5% and 12.0% accuracy, respectively. The model also struggled significantly with Reasoning, scoring 0.0% accuracy, and Coding, with only 2.0% accuracy. While Email Classification reached 94.0% accuracy, its percentile ranking (29th) suggests it is not a top performer in this specific task. Instruction Following also presented a challenge, with 30.0% accuracy. Overall, Lumimaid v0.2 8B's primary strengths lie in its speed and cost-efficiency. Its significant weaknesses are in accuracy across a broad range of cognitive tasks, particularly reasoning, general knowledge, and ethical judgment. The model appears to be a strong candidate for applications where speed and low cost are paramount, and accuracy requirements are either less stringent or can be augmented by other systems.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.15
Completion $1.5

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Mancer 2
Mancer 2 | neversleep/llama-3.1-lumimaid-8b 32K $0.15 / 1M tokens $1.5 / 1M tokens
Featherless
Featherless | neversleep/llama-3.1-lumimaid-8b 16K $0.09 / 1M tokens $0.6 / 1M tokens
NextBit
NextBit | neversleep/llama-3.1-lumimaid-8b 32K $0.09 / 1M tokens $0.6 / 1M tokens
Parasail
Parasail | neversleep/llama-3.1-lumimaid-8b 40K $0.09 / 1M tokens $0.6 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by neversleep