newsbang

Homer-v1.0-Qwen2.5-72B

fine-tunedondomain-specificdatasetsQwen2ForCausalLMbfloat16

Install and run this model locally using llmpm, the open-source LLM package manager.

Install
llmpm install newsbang/Homer-v1.0-Qwen2.5-72B
Run
llmpm run newsbang/Homer-v1.0-Qwen2.5-72B
Average Score (0–100)
47.5%
Weighted average of normalized scores from all benchmarks. Each benchmark is normalized to a 0–100 scale, then averaged together.

BENCHMARK SCORES

IFEval76.3%

Instruction-Following Evaluation. Tests the model's ability to follow explicit formatting instructions (instruction following, formatting, generation). Scored by strict format accuracy.

BBH62.3%

Big Bench Hard. A collection of challenging tasks across language understanding, mathematical reasoning, and common sense knowledge. Scored by accuracy on multiple-choice questions.

MATH Lvl 549.0%

Mathematics Aptitude Test of Heuristics, Level 5. High school competition problems covering complex algebra, geometry, and advanced calculus. Scored by exact match.

GPQA22.1%

Graduate-Level Google-Proof Q&A. PhD-level multiple-choice questions in chemistry, biology, and physics. Scored by accuracy.

MuSR17.9%

Multistep Soft Reasoning. Tests reasoning and understanding over long texts, including language understanding, reasoning capabilities, and long-context reasoning. Scored by accuracy.

MMLU-Pro57.2%

Massive Multitask Language Understanding – Professional. Expert-reviewed multiple-choice questions across medicine, law, engineering, and mathematics. Scored by accuracy.

MODEL INFO

Architecture
Qwen2ForCausalLM
Precision
bfloat16
Type
fine-tunedondomain-specificdatasets
Weight Type
Original
Parameters
72.7B
Chat Template
No

METADATA

Upload Date
2024-12-16
Submission Date
2024-12-16
License
apache-2.0
Base Model
newsbang/Homer-v1.0-Qwen2.5-72B
HF Hearts
6
CO₂ Cost (kg)
29.55