Skip to content

Commit 7c2a322

Browse files
committed
added an image and fixed some linking
1 parent 4036bb9 commit 7c2a322

File tree

5 files changed

+10
-10
lines changed

5 files changed

+10
-10
lines changed

guides/deleteinstance.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Stopping Instances helps save money on GPU deployments when you're not using the
1010
### Setup
1111
Let's assume that you have a Shadeform API key from [here](https://platform.shadeform.ai/settings/api),
1212
and an active instance. If you need an active instance, then
13-
[this guide](https://docs.shadeform.ai/findingmostaffordablegpus) will help you get set up.
13+
[this guide](/guides/mostaffordablegpus) will help you get set up.
1414

1515
### Deleting an instance:
1616

guides/mostaffordablegpus.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ In such cases, we provide an API that can help search for the most affordable GP
1515

1616
### Using the Instance Types API
1717

18-
We can find the most affordable GPU instances by applying filtering and sorting query params to the [Instance Types API](/api-reference/endpoint/instances-types).
18+
We can find the most affordable GPU instances by applying filtering and sorting query params to the [Instance Types API](/api-reference/instances/instances-types).
1919
Without any query parameters, the API returns all of the instance types supported by Shadeform in a random order.
2020
By using the following query parameters, we can filter down the results and sort them as needed.
2121

@@ -259,7 +259,7 @@ From array of results returned, we can see that the A6000 instance from RunPod i
259259
```
260260

261261
### Launching the Instance
262-
After you have found the most affordable instance, you can launch the instance by using the [Create Instance API](/api-reference/endpoint/instances-create).
262+
After you have found the most affordable instance, you can launch the instance by using the [Create Instance API](/api-reference/instances/instances-create).
263263
In the example below, we will launch the most affordable A6000 GPU based on the data retrieved from the API call in the previous section.
264264
We will take the following properties from the previous response:
265265

guides/serveembeddings.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,9 +12,9 @@ In this guide, we will show you how you can serve an embedding model using 'Text
1212
This framework runs inference very quicky and can handle large batch sizes.
1313

1414
### Setup
15-
This guide builds off of our others for [finding the best gpu](https://docs.shadeform.ai/findingmostaffordablegpus).
15+
This guide builds off of our others for [finding the best gpu](/guides/mostaffordablegpus).
1616

17-
We have a python notebook already to go for you to deploy this model that you can [find here](https://github.com/shadeform/examples/blob/main/basic_serving_embeddigns_tei.ipynb).
17+
We have a python notebook already to go for you to deploy this model that you can [find here](https://github.com/shadeform/examples/blob/main/serving_embeddings_tei.ipynb).
1818
The requirements are simple, so in a python environment with `requests` installed:
1919

2020
```bash
@@ -71,7 +71,7 @@ print(response.text)
7171

7272
Once we request it, Shadeform will provision the machine, and deploy a docker container based on the image, arguments, and environment variables that we selected.
7373
This might take 5-10 minutes depending on the machine chosen and the size of the model weights you choose.
74-
For more information on the API fields, check out the [Create Instance API Reference](https://docs.shadeform.ai/api-reference/endpoint/instances-create).
74+
For more information on the API fields, check out the [Create Instance API Reference](/api-reference/instances/instances-create).
7575

7676
### Checking on our server
7777

guides/tgi.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Model Deployment is a very common GPU use case. [Text-Generation-Inference](http
1010

1111
With Shadeform, it's easy to deploy models right to the most affordable gpu's in the market with just a few commands.
1212

13-
In this guide, we will deploy [Mistral-7b-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) with `TGI` onto an A6000. This guide is very similar to our [vLLM guide]() with a few changes to change the inference framework.
13+
In this guide, we will deploy [Mistral-7b-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) with `TGI` onto an A6000. This guide is very similar to our [vLLM guide](/guides/vllm) with a few changes to change the inference framework.
1414

1515
```bash
1616
git clone https://github.com/shadeform/examples.git
@@ -69,7 +69,7 @@ print(response.text)
6969
Once we request it, Shadeform will provision the machine, and deploy a docker container with the image, arguments, and environment variables provided. In this case, it will deploy an openai compatible server with `TGI` serving Mistral-7b-v0.1.
7070
This might take 5-10 minutes depending on the machine chosen and the size of the model weights you choose.
7171

72-
For more information on the API fields, check out the [Create Instance API Reference](https://docs.shadeform.ai/api-reference/endpoint/instances-create).
72+
For more information on the API fields, check out the [Create Instance API Reference](/api-reference/instances/instances-create).
7373

7474
### Checking on our Model server
7575

guides/vllm.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ In this guide, we will deploy [Mistral-7b-v0.1](https://huggingface.co/mistralai
1111

1212
### Setup
1313

14-
This guide builds off of our others for [finding the best gpu](https://docs.shadeform.ai/findingmostaffordablegpus) and for [deploying gpu containers](https://docs.shadeform.ai/deployinggpucontainers).
14+
This guide builds off of our others for [finding the best gpu](/guides/mostaffordablegpus) and for [deploying gpu containers](/guides/dockercontainers).
1515
We have a python notebook already to go for you to deploy this model that you can [find here](https://github.com/shadeform/examples/blob/main/basic_serving_vllm.ipynb).
1616
The requirements are simple, so in a python environment with (`requests` + optionally `openai`) installed:
1717

@@ -70,7 +70,7 @@ print(response.text)
7070
Once we request it, Shadeform will provision the machine, and deploy a docker container with the image, arguments, and environment variables provided. In this case, it will deploy an openai compatible server with `vLLM` serving Mistral-7b-v0.1.
7171
This might take 5-10 minutes depending on the machine chosen and the size of the model weights you choose.
7272

73-
For more information on the API fields, check out the [Create Instance API Reference](https://docs.shadeform.ai/api-reference/endpoint/instances-create).
73+
For more information on the API fields, check out the [Create Instance API Reference](/api-reference/instances/instances-create).
7474

7575
### Checking on our Model server
7676

0 commit comments

Comments
 (0)