STEM-en-ms / README.md
Supa-AI's picture
Update README.md
505dd13 verified
|
raw
history blame
8.87 kB
---
license: cc-by-4.0
dataset_info:
- config_name: data-ms
features:
- name: file name
dtype: string
- name: IBSN
dtype: string
- name: subject
dtype: string
- name: topic
dtype: string
- name: Questions
dtype: string
- name: figures
sequence: image
- name: label
sequence: string
- name: Options
dtype: string
- name: Answers
dtype: string
splits:
- name: eval
num_bytes: 34663548
num_examples: 614
download_size: 34559856
dataset_size: 34663548
- config_name: data_en
features:
- name: FileName
dtype: string
- name: IBSN
dtype: string
- name: Subject
dtype: string
- name: Topic
dtype: string
- name: Questions
dtype: string
- name: Figures
sequence: image
- name: Label
sequence: string
- name: Options
dtype: string
- name: Answers
dtype: string
splits:
- name: eval
num_bytes: 34663548
num_examples: 614
download_size: 69119656
dataset_size: 69327096.0
tags:
- mathematics
- physics
- llms
- Malaysia
- Asia
size_categories:
- n<1K
configs:
- config_name: data_en
data_files:
- split: eval
path: data_en/train-*
- config_name: data_ms
data_files:
- split: eval
path: data_ms/train-*
language:
- en
- ms
---
# **A Bilingual Dataset for Evaluating Reasoning Skills in STEM Subjects**
This dataset provides a comprehensive evaluation set for tasks assessing reasoning skills in Science, Technology, Engineering, and Mathematics (STEM) subjects. It features questions in both English and Malay, catering to a diverse audience.
**Key Features**
* **Bilingual:** Questions are available in English and Malay, promoting accessibility for multilingual learners.
* **Visually Rich:** Questions are accompanied by figures to enhance understanding and support visual and contextual reasoning.
* **Focus on Reasoning:** The dataset emphasizes questions requiring logical reasoning and problem-solving skills, as opposed to simple recall of knowledge.
* **Real-World Context:** Questions are derived from real-world scenarios, such as past SPM (Sijil Pelajaran Malaysia) examinations, making them relatable to students.
**Dataset Structure**
The dataset is comprised of two configurations: `data_en` (English) and `data_ms` (Malay). Both configurations share the same features and structure.
**Data Fields**
* **FileName:** Unique identifier for the source file (alphanumeric).
* **IBSN:** International Standard Book Number of the source book (if available).
* **Subject:** Academic subject (e.g., Physics, Mathematics).
* **Topic:** Specific topic of the question within the subject (may be missing).
* **Questions:** Main body of the question or problem statement.
* **Figures:** List of associated image files related to the question (empty if no figures are present).
* **Label:** Original caption or description of each image in the `imgs` list.
* **Options:** Possible answer choices for the question, with keys (e.g., "A", "B", "C", "D") and corresponding text.
* **Answers:** Correct answer to the question, represented by the key of the correct option (e.g., "C").
---
## Data Instance Example
```json
{
    "FileName": "FC064244",
    "ISBN": "9786294703681",
    "Subject": "Physics",
    "Topic": "Measurement",
    "Questions": "State the physical quantity that can be measured using the measuring device shown in Diagram 1.",
    "Figures": [
        {
            "label": "Diagram 1",
            "path": "FC064244_C1_Q12_ImageFile_0.png"
        }
    ],
    "Options": {
        "A": "Weight",
        "B": "Mass",
        "C": "Amount of substance",
        "D": "Volume"
    },
    "Answers": "B"
}
```
**Data Split**
The dataset is split between Physics and Mathematics subjects, with some questions lacking topic categorization.
| Subject     | Instances with Topic | Instances without Topic | Total |
|-------------|----------------------|-------------------------|-------|
| Physics     | 316                  | 77                      | 393   |
| Mathematics | 32                   | 189                     | 221   |
**Known Limitations**
* **Subject Coverage:** The current version focuses on Physics and Mathematics. Future releases will include more STEM subjects.
* **Answer Accuracy:** Answers are extracted from various sources and may contain inaccuracies.
**Source**
The dataset is derived from a combination of resources, including:
* SPM past-year exams
* SPM mock exams
* Educational exercise books
**Data Acquisition Method**
* Optical Character Recognition (OCR) for text extraction
* Manual quality control (QC) to ensure data accuracy
**Versioning and Maintenance**
* **Current Version:** 1.0.0
* **Release Date:** December 27, 2024
* **Contact:** We welcome any feedback or corrections to improve the dataset quality.
---
# License
This dataset is licensed under the [Creative Commons Attribution 4.0 International License (CC BY 4.0)](https://creativecommons.org/licenses/by/4.0/).
---
# Getting Started
You can access the dataset on Hugging Face using the following commands:
```bash
# For English data
pip install datasets
from datasets import load_dataset
dataset = load_dataset("Supa-AI/STEM-en-ms", name="data_en")
# For Malay data
dataset = load_dataset("Supa-AI/STEM-en-ms", name="data_ms")
```
---
# Bilingual STEM Dataset LLM Leaderboard
This document summarizes the evaluation results for various language models based on **5-shot** and **First Token Accuracy**. The evaluation was conducted across four configurations:
| **Model** | **en\_withfigures** | **en\_withoutfigures** | **ms\_withfigures** | **ms\_withoutfigures** |
| --------------------------------- | ------------------- | ---------------------- | ------------------- | ---------------------- |
| **gemini-2.0-flash-exp** | **63.70%** | <ins>75.16%</ins> | **63.36%** | <ins>75.47%</ins> |
| **gemini-1.5-flash** | 49.66% | 67.39% | 50.00% | 64.28% |
| **Qwen/Qwen2-VL-72B-Instruct** | <ins>58.22%</ins> | 69.25% | <ins>57.53%</ins> | 63.66% |
| **gpt-4o** | 47.95% | 66.15% | 50.00% | 68.01% |
| **gpt-4o-mini** | 41.10% | 55.90% | 38.36% | 52.80% |
| **pixtral-large-2411** | 42.81% | 64.29% | 35.27% | 60.87% |
| **pixtral-12b-2409** | 24.66% | 48.45% | 24.66% | 39.13% |
| **DeepSeek-V3** | None | **79.19%** | None | **76.40%** |
| **Qwen2.5-72B-Instruct** | None | 74.53% | None | 72.98% |
| **Meta-Llama-3.3-70B-Instruct** | None | 67.08% | None | 58.07% |
| **Llama-3.2-90B-Vision-Instruct** | None | 65.22% | None | 58.07% |
| **sail/Sailor2-20B-Chat** | None | 66.46% | None | 61.68% |
| **mallam-small** | None | 61.49% | None | 55.28% |
| **mistral-large-latest** | None | 60.56% | None | 53.42% |
| **google/gemma-2-27b-it** | None | 58.07% | None | 57.76% |
| **SeaLLMs-v3-7B-Chat** | None | 50.93% | None | 45.96% |
---
## Notes
In the repository, there is an `eval.py` script that can be used to run the evaluation for any other LLM.
The evaluation results are based on the specific dataset and methodology employed.
- The "First Token Accuracy" metric emphasizes the accuracy of predicting the initial token correctly.
- Further analysis might be needed to determine the models' suitability for specific tasks.
### Attribution for Evaluation Code
The `eval.py` script is based on work from the MMLU-Pro repository:
- Repository: [TIGER-AI-Lab/MMLU-Pro](https://github.com/TIGER-AI-Lab/MMLU-Pro)
- License: Apache License 2.0 (included in the `NOTICE` file)
---
# **Contributors**
- [**Gele**](https://huggingface.co/Geleliong)
- [**Ken Boon**](https://huggingface.co/caibcai)
- [**Wei Wen**](https://huggingface.co/WeiWen21)