Fireworks AI
Fireworks AI offers a rapid, cost-effective, and adaptable solution for generative artificial intelligence, enabling product developers to execute, refine, and distribute LLMs efficiently.
Deploying your model​
Fireworks offers the capability to deploy various Chat Models(LLM) under its umbrella.
-
Environment Variable :​
Below value(s) are to be added in
.env
file, present atbuilder/partnerproduct/
.FIREWORKS_API_KEY : <check_references_below>
Chat Model​
Refer to their documentation to understand the latest offerings, with feature and cost comparisons.
Usage with MAAP​
To use Fireworks model with MAAP framework, you would need to feed below values.
-
Config File :​
Provided below are the values required to be added inconfig.yaml
file in embedding section.llms:
class_name: Fireworks
model_name: <check_references_below>
max_tokens: <integer_value>
temperature: <integer_value>
Embedding Model​
You can follow the same steps as above to deploy the embedding model as well. The process is documented here.
Usage with MAAP​
To use FireworksAI embedding with MAAP framework, you would need to feed below values.
-
Config File :​
Provided below are the values required to be added inconfig.yaml
file in embedding section. Theclass_name
should be appropriate to the embedding model being used.embedding:
class_name: Fireworks
model_name: <check_references_below>
Deploying your model using the LlamaIndex framework​
Chat Model​
MAAP now provides the option to choose if you want to use LlamaIndex as your main framework to deploy your LLM models.
This can be done by adding the 'framework' configuration to the config.yaml file
-
Config File​
llms:
class_name: Fireworks
model_name: <check_references_below>
framework: 'llamaindex'
Embedding Model​
MAAP now provides the option to choose if you want to use LlamaIndex as your main framework to deploy your embeddings.
This can be done by adding the 'framework' configuration to the config.yaml file
-
Config File​
Another important consideration to have is that because of the way that LlamaIndex implements its embedding models, the environment variables that you are using cannot contain any parameters related to AzureOpenAI. This is because LlamaIndex automatically detects if you have any Azure environment variables and uses their endpoints if they are set.embedding:
class_name: Fireworks
model_name: <check_references_below>
framework: 'llamaindex'
References​
Provided below are the instructions on how to procure the right values for building your MAAP framework.
-
Chat Model Name​
You can pick any model from the updated LLM model list given in Fireworks documentation.
-
Embedding Model Name​
You can pick any model from the updated embedding model list given in Fireworks documentation.
-
API Key​
You will need to sign up and retrieve a Fireworks API Key.