Skip to content

Commit

Permalink
Init
Browse files Browse the repository at this point in the history
  • Loading branch information
dmatora committed May 2, 2024
0 parents commit 055772a
Show file tree
Hide file tree
Showing 3 changed files with 92 additions and 0 deletions.
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
.idea
21 changes: 21 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,21 @@
# LLM Inference Speeds

This repository contains benchmark data for various Large Language Models (LLM) based on their inference speeds measured in tokens per second. The benchmarks are performed across different hardware configurations using the prompt "tell a story".

## About the Data

The data represents the performance of several LLMs, detailing the tokens processed per second on specific hardware setups. Each entry includes the model name, the hardware used, and the measured speed.

## Explore the Benchmarks

You can view and interact with the benchmark data through a searchable table on our GitHub Pages site. Use the search field to filter by model name and explore different hardware performances.

**[View the Inference Speeds Table](https://dmatora.github.io/inference-speed/)**

## Contributing

Contributions to the benchmark data are welcome! Please refer to the contributing guidelines for more information on how you can contribute.

## License

This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
70 changes: 70 additions & 0 deletions index.html
Original file line number Diff line number Diff line change
@@ -0,0 +1,70 @@
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>LLM Inference Speeds</title>
<style>
body { font-family: Arial, sans-serif; }
input { margin: 20px 0; padding: 10px; width: 200px; }
table { width: 100%; border-collapse: collapse; }
th, td { padding: 10px; border: 1px solid #ddd; text-align: left; }
th { background-color: #f2f2f2; }
</style>
</head>
<body>
<h1>LLM Inference Speeds</h1>
<input type="text" id="searchInput" onkeyup="filterModels()" placeholder="Search for models...">

<table id="dataTable">
<thead>
<tr>
<th>Model</th>
<th>Hardware</th>
<th>Speed</th>
</tr>
</thead>
<tbody>
</tbody>
</table>

<script>
document.addEventListener('DOMContentLoaded', function() {
fetch('data.json')
.then(response => response.json())
.then(data => {
const tableBody = document.getElementById('dataTable').getElementsByTagName('tbody')[0];
data.forEach(item => {
let row = tableBody.insertRow();
let cell1 = row.insertCell(0);
let cell2 = row.insertCell(1);
let cell3 = row.insertCell(2);
cell1.textContent = item.model;
cell2.textContent = item.hardware;
cell3.textContent = item.speed;
});
})
.catch(error => console.error('Error loading the data:', error));
});

function filterModels() {
var input, filter, table, tr, td, i, txtValue;
input = document.getElementById("searchInput");
filter = input.value.toUpperCase();
table = document.getElementById("dataTable");
tr = table.getElementsByTagName("tr");
for (i = 0; i < tr.length; i++) {
td = tr[i].getElementsByTagName("td")[0];
if (td) {
txtValue = td.textContent || td.innerText;
if (txtValue.toUpperCase().indexOf(filter) > -1) {
tr[i].style.display = "";
} else {
tr[i].style.display = "none";
}
}
}
}
</script>
</body>
</html>

0 comments on commit 055772a

Please sign in to comment.