Skip to content

Commit 5ef516c

Browse files
authored
Merge pull request #23 from Sherlock113/docs/seo-optimization
docs: Optimize SEO
2 parents 1c047a4 + 7632ce4 commit 5ef516c

31 files changed

+77
-60
lines changed

docs/getting-started/calculating-gpu-memory-for-llms.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
sidebar_position: 2
33
description: Learn how to calculate GPU memory for serving LLMs.
44
keywords:
5-
- GPU memory calculation
5+
- GPU memory calculation, LLM inference hardware calculator
66
- VRAM calculation
77
- LLM memory requirements
88
---

docs/getting-started/choosing-the-right-inference-framework.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,9 +2,8 @@
22
sidebar_position: 5
33
description: Select the right inference frameworks for your use case.
44
keywords:
5-
- Inference frameworks
6-
- Inference backends
7-
- Inference runtimes
5+
- Inference frameworks, inference backends, inference runtimes, inference engines, inference platforms
6+
- Best inference frameworks, best LLM inference providers, LLM inference benchmark
87
- vLLM, SGLang, LMDeploy, TensorRT-LLM, Hugging Face TGI, llama.cpp, MLC-LLM, Ollama
98
---
109

docs/getting-started/choosing-the-right-model.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
sidebar_position: 1
33
description: Select the right models for your use case.
44
keywords:
5-
- LLMs
5+
- LLMs, dense models
66
- Base models
77
- Instruction-tuned models
88
- Mixture of Experts models

docs/getting-started/llm-fine-tuning.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
---
22
sidebar_position: 3
3-
description: Understand LLM fine-tuning and different fine-tuning frameworks
3+
description: Understand LLM fine-tuning and different fine-tuning frameworks.
44
keywords:
5-
- LLM fine-tuning
6-
- Fine-tuning frameworks
5+
- LLM fine-tuning, LoRA, how does LLM fine-tuning work
6+
- Fine-tuning frameworks, open source LLM fine-tuning, types of LLM fine-tuning
77
- Axolotl, Unsloth, Torchtune, LLaMA Factory
88
---
99

docs/getting-started/llm-quantization.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
---
22
sidebar_position: 4
3-
description: Understand LLM quantization and different quantization formats and methods
3+
description: Understand LLM quantization and different quantization formats and methods.
44
keywords:
5-
- LLM quantization
6-
- Quantization formats
5+
- LLM quantization, how does quantization work, LLM quantization accuracy
6+
- Quantization formats, quantization types, quantization techniques
77
- AWQ, SmoothQuant, GPTQ
88
---
99

docs/getting-started/tool-integration/_category_.json

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
{
22
"label": "Tool integration",
33
"position": 6,
4+
"collapsed": false,
45
"link": {
56
"type": "generated-index"
67
}

docs/getting-started/tool-integration/function-calling.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,8 @@
22
sidebar_position: 1
33
description: Learn what function calling is and its use case.
44
keywords:
5-
- Function calling
5+
- Function calling, function calling APIs
6+
- LLM tool use, LLM tool integration
67
---
78

89
import LinkList from '@site/src/components/LinkList';

docs/getting-started/tool-integration/model-context-protocol.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,8 @@
22
sidebar_position: 2
33
description: Learn what Model Context Protocol (MCP) is and its use case.
44
keywords:
5-
- Model Context Protocol
6-
- MCP
5+
- Model Context Protocol, MCP protocol
6+
- MCP host, MCP clients, MCP servers
77
---
88

99
# Model Context Protocol

docs/inference-optimization/data-tensor-pipeline-expert-hybrid-parallelism.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,9 @@
22
sidebar_position: 9
33
description: Understand the differences between data, tensor, pipeline, expert and hybrid parallelisms.
44
keywords:
5-
- LLM inference optimization
5+
- LLM inference optimization, LLM inference optimization techniques​
66
- Data parallelism, tensor parallelism, pipeline parallelism, expert parallelism and hybrid parallelism
7+
- Speed up LLM inference
78
---
89

910
import LinkList from '@site/src/components/LinkList';

docs/inference-optimization/kv-cache-utilization-aware-load-balancing.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,9 +3,10 @@ sidebar_position: 8
33
description: Route LLM requests based on KV cache usage for faster, smarter inference.
44
keywords:
55
- KV cache
6-
- Load balancing
7-
- LLM inference optimization
6+
- Load balancing, LLM load balancing
7+
- LLM inference optimization, LLM inference optimization techniques​
88
- Gateway API Inference Extension
9+
- Speed up LLM inference
910
---
1011

1112
# KV cache utilization-aware load balancing

0 commit comments

Comments
 (0)