Author's Description
Lumimaid v0.2 8B is a finetune of [Llama 3.1 8B](/models/meta-llama/llama-3.1-8b-instruct) with a "HUGE step up dataset wise" compared to Lumimaid v0.1. Sloppy chats output were purged. Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).
Key Specifications
Supported Parameters
This model supports the following parameters:
Features
This model supports the following features:
Performance Summary
NeverSleep: Lumimaid v0.2 8B demonstrates exceptional performance in terms of operational efficiency, consistently ranking among the fastest models and offering highly competitive pricing across all benchmarks. This makes it a cost-effective and responsive option for various applications. However, its accuracy across most benchmark categories is notably low. In Hallucinations (Baseline), it achieved 60.0% accuracy, indicating a tendency to generate information without acknowledging uncertainty. General Knowledge and Ethics benchmarks showed particularly weak performance at 20.5% and 12.0% accuracy, respectively. The model also struggled significantly with Reasoning, scoring 0.0% accuracy, and Coding, with only 2.0% accuracy. While Email Classification reached 94.0% accuracy, its percentile ranking (29th) suggests it is not a top performer in this specific task. Instruction Following also presented a challenge, with 30.0% accuracy. Overall, Lumimaid v0.2 8B's primary strengths lie in its speed and cost-efficiency. Its significant weaknesses are in accuracy across a broad range of cognitive tasks, particularly reasoning, general knowledge, and ethical judgment. The model appears to be a strong candidate for applications where speed and low cost are paramount, and accuracy requirements are either less stringent or can be augmented by other systems.
Model Pricing
Current Pricing
Feature | Price (per 1M tokens) |
---|---|
Prompt | $0.15 |
Completion | $1.5 |
Price History
Available Endpoints
Provider | Endpoint Name | Context Length | Pricing (Input) | Pricing (Output) |
---|---|---|---|---|
Mancer 2
|
Mancer 2 | neversleep/llama-3.1-lumimaid-8b | 32K | $0.15 / 1M tokens | $1.5 / 1M tokens |
Featherless
|
Featherless | neversleep/llama-3.1-lumimaid-8b | 16K | $0.09 / 1M tokens | $0.6 / 1M tokens |
NextBit
|
NextBit | neversleep/llama-3.1-lumimaid-8b | 32K | $0.09 / 1M tokens | $0.6 / 1M tokens |
Parasail
|
Parasail | neversleep/llama-3.1-lumimaid-8b | 40K | $0.09 / 1M tokens | $0.6 / 1M tokens |
Benchmark Results
Benchmark | Category | Reasoning | Strategy | Free | Executions | Accuracy | Cost | Duration |
---|
Other Models by neversleep
|
Released | Params | Context |
|
Speed | Ability | Cost |
---|---|---|---|---|---|---|---|
NeverSleep: Lumimaid v0.2 70B Unavailable | Oct 21, 2024 | 70B | 16K |
Text input
Text output
|
★★ | ★★★★★ | $$$$$ |
NeverSleep: Llama 3 Lumimaid 70B Unavailable | May 15, 2024 | 70B | 8K |
Text input
Text output
|
★ | ★ | $$$$$ |
NeverSleep: Llama 3 Lumimaid 8B Unavailable | May 03, 2024 | 8B | 24K |
Text input
Text output
|
★★★★ | ★★ | $$$ |
Noromaid 20B | Nov 25, 2023 | 20B | 8K |
Text input
Text output
|
★★★ | ★ | $$$$ |