Skip to content

Commit 636f87f

Browse files
docs: faq style change, toc added to index
docs: updated style in FAQ added: TOC to the index page
2 parents 01599ca + 0733813 commit 636f87f

File tree

2 files changed

+39
-33
lines changed

2 files changed

+39
-33
lines changed

docs/docs/FAQ.md

Lines changed: 24 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
# Frequently Asked Questions (FAQ)
22

3-
## What are the best models for bolt.diy?
3+
<details>
4+
<summary><strong>What are the best models for bolt.diy?</strong></summary>
45

56
For the best experience with bolt.diy, we recommend using the following models:
67

@@ -11,10 +12,10 @@ For the best experience with bolt.diy, we recommend using the following models:
1112
- **Qwen 2.5 Coder 32b**: Best model for self-hosting with reasonable hardware requirements
1213

1314
**Note**: Models with less than 7b parameters typically lack the capability to properly interact with bolt!
15+
</details>
1416

15-
---
16-
17-
## How do I get the best results with bolt.diy?
17+
<details>
18+
<summary><strong>How do I get the best results with bolt.diy?</strong></summary>
1819

1920
- **Be specific about your stack**:
2021
Mention the frameworks or libraries you want to use (e.g., Astro, Tailwind, ShadCN) in your initial prompt. This ensures that bolt.diy scaffolds the project according to your preferences.
@@ -28,72 +29,62 @@ For the best experience with bolt.diy, we recommend using the following models:
2829
- **Batch simple instructions**:
2930
Combine simple tasks into a single prompt to save time and reduce API credit consumption. For example:
3031
*"Change the color scheme, add mobile responsiveness, and restart the dev server."*
32+
</details>
3133

32-
---
33-
34-
## How do I contribute to bolt.diy?
34+
<details>
35+
<summary><strong>How do I contribute to bolt.diy?</strong></summary>
3536

3637
Check out our [Contribution Guide](CONTRIBUTING.md) for more details on how to get involved!
38+
</details>
3739

38-
---
39-
40-
## What are the future plans for bolt.diy?
40+
<details>
41+
<summary><strong>What are the future plans for bolt.diy?</strong></summary>
4142

4243
Visit our [Roadmap](https://roadmap.sh/r/ottodev-roadmap-2ovzo) for the latest updates.
4344
New features and improvements are on the way!
45+
</details>
4446

45-
---
46-
47-
## Why are there so many open issues/pull requests?
47+
<details>
48+
<summary><strong>Why are there so many open issues/pull requests?</strong></summary>
4849

4950
bolt.diy began as a small showcase project on @ColeMedin's YouTube channel to explore editing open-source projects with local LLMs. However, it quickly grew into a massive community effort!
5051

51-
We’re forming a team of maintainers to manage demand and streamline issue resolution. The maintainers are rockstars, and we’re also exploring partnerships to help the project thrive.
52-
53-
---
52+
We're forming a team of maintainers to manage demand and streamline issue resolution. The maintainers are rockstars, and we're also exploring partnerships to help the project thrive.
53+
</details>
5454

55-
## How do local LLMs compare to larger models like Claude 3.5 Sonnet for bolt.diy?
55+
<details>
56+
<summary><strong>How do local LLMs compare to larger models like Claude 3.5 Sonnet for bolt.diy?</strong></summary>
5657

5758
While local LLMs are improving rapidly, larger models like GPT-4o, Claude 3.5 Sonnet, and DeepSeek Coder V2 236b still offer the best results for complex applications. Our ongoing focus is to improve prompts, agents, and the platform to better support smaller local LLMs.
59+
</details>
5860

59-
---
60-
61-
## Common Errors and Troubleshooting
61+
<details>
62+
<summary><strong>Common Errors and Troubleshooting</strong></summary>
6263

6364
### **"There was an error processing this request"**
6465
This generic error message means something went wrong. Check both:
6566
- The terminal (if you started the app with Docker or `pnpm`).
6667
- The developer console in your browser (press `F12` or right-click > *Inspect*, then go to the *Console* tab).
6768

68-
---
69-
7069
### **"x-api-key header missing"**
7170
This error is sometimes resolved by restarting the Docker container.
72-
If that doesn’t work, try switching from Docker to `pnpm` or vice versa. We’re actively investigating this issue.
73-
74-
---
71+
If that doesn't work, try switching from Docker to `pnpm` or vice versa. We're actively investigating this issue.
7572

7673
### **Blank preview when running the app**
7774
A blank preview often occurs due to hallucinated bad code or incorrect commands.
7875
To troubleshoot:
7976
- Check the developer console for errors.
80-
- Remember, previews are core functionality, so the app isn’t broken! We’re working on making these errors more transparent.
81-
82-
---
77+
- Remember, previews are core functionality, so the app isn't broken! We're working on making these errors more transparent.
8378

8479
### **"Everything works, but the results are bad"**
8580
Local LLMs like Qwen-2.5-Coder are powerful for small applications but still experimental for larger projects. For better results, consider using larger models like GPT-4o, Claude 3.5 Sonnet, or DeepSeek Coder V2 236b.
8681

87-
---
88-
8982
### **"Received structured exception #0xc0000005: access violation"**
90-
9183
If you are getting this, you are probably on Windows. The fix is generally to update the [Visual C++ Redistributable](https://learn.microsoft.com/en-us/cpp/windows/latest-supported-vc-redist?view=msvc-170)
9284

93-
---
94-
9585
### **"Miniflare or Wrangler errors in Windows"**
9686
You will need to make sure you have the latest version of Visual Studio C++ installed (14.40.33816), more information here https://github.com/stackblitz-labs/bolt.diy/issues/19.
87+
</details>
9788

9889
---
9990

docs/docs/index.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,21 @@
11
# Welcome to bolt diy
22
bolt.diy allows you to choose the LLM that you use for each prompt! Currently, you can use OpenAI, Anthropic, Ollama, OpenRouter, Gemini, LMStudio, Mistral, xAI, HuggingFace, DeepSeek, or Groq models - and it is easily extended to use any other model supported by the Vercel AI SDK! See the instructions below for running this locally and extending it to include more models.
33

4+
## Table of Contents
5+
- [Join the community!](#join-the-community)
6+
- [What's bolt.diy](#whats-boltdiy)
7+
- [What Makes bolt.diy Different](#what-makes-boltdiy-different)
8+
- [Setup](#setup)
9+
- [Run with Docker](#run-with-docker)
10+
- [Using Helper Scripts](#1a-using-helper-scripts)
11+
- [Direct Docker Build Commands](#1b-direct-docker-build-commands-alternative-to-using-npm-scripts)
12+
- [Docker Compose with Profiles](#2-docker-compose-with-profiles-to-run-the-container)
13+
- [Run Without Docker](#run-without-docker)
14+
- [Adding New LLMs](#adding-new-llms)
15+
- [Available Scripts](#available-scripts)
16+
- [Development](#development)
17+
- [Tips and Tricks](#tips-and-tricks)
18+
419
---
520

621
## Join the community!

0 commit comments

Comments
 (0)