Skip to main content

Cohere

Cohere Inc. is a Canadian multinational technology firm that concentrates on enterprise artificial intelligence, specializing particularly in large-scale language models.

Deploying your model​

Cohere offers the capability to deploy various Chat Models(LLM) under its umbrella.

Chat Model​

Refer to their documentation to understand the latest offerings, with feature and cost comparisons.

Usage with MAAP​

To use the Cohere chat model with the MAAP framework, you would need to feed the below values.

  • Config File :​

    Provided below are the values required to be added in the config.yaml file in the LLM section.

    llms:
    class_name: Cohere
    model_name: <check_references_below>
    temperature: <integer_value>
  • Environment Variable :​

    The below value(s) are to be added in the .env file, present at builder/partnerproduct/.

    COHERE_API_KEY : <check_references_below>

Embedding Model​

Refer to their documentation to understand the latest offerings, with feature and cost comparisons.

Usage with MAAP​

To use the Cohere embedding model with the MAAP framework, you would need to feed the below values.

  • Config File :​

    Provided below are the values required to be added in the config.yaml file located in the embedding section.

    embedding:
    class_name: Cohere
    model_name: <check_references_below>
  • Environment Variable :​

    The below value(s) are to be added in the .env file, present at builder/partnerproduct/.

    COHERE_API_KEY = <check_references_below>

Deploying your model using the LlamaIndex framework​

MAAP now provides the option to choose if you want to use LlamaIndex as your main framework to deploy your LLM models and embeddings.

This can be done by adding the 'framework' configuration to the config.yaml file

Chat Model​

  • Config File​

    llms:
    class_name: Cohere
    model_name: <check_references_below>
    framework: 'LlamaIndex'

Embedding Model​

  • Config File​

    embedding:
    class_name: Cohere
    model_name: <check_references_below>
    framework: 'LlamaIndex'

For optimal quality the recommended length of each chunk should be under 512 tokens.

References​

Provided below are the instructions on how to procure the right values for building your MAAP framework.

For more information about how it was implemented for LlamaIndex click here.