Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

What are the LLMs that can be used?

Any provisioned models can be connected to the system via config. So far all, OpenAI Models have been tested:

  • GPT-3.5 - GPT-3.5-Turbo

  • GPT-4 - GPT 4-Turbo

  • GPT-4 Vision (in various minor versions)

  • GPT-4o

Other LLM’s:

  • Mistral AI - Zephyr AI

Is the availability of Azure OpenAI models restricted to certain regions?

Yes, the availability of Azure OpenAI models is dependent on the deployment region. The specific models available in each region can be checked on the Azure website:

https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models

This page is regularly updated, allowing users to see when new models become available in different regions.

Will Azure retire OpenAI models over time?

Yes, Microsoft does retire OpenAI models over time. Information about model retirements is provided on the Azure page under “Model Retirements.”

https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/model-retirements

This section includes a table listing the current models, their versions, and the retirement dates. Clients will be informed in advance about any model retirements, as well as new models becoming available in their region.

What is the difference between model retirement and deprecation communicated by Azure for their provided OpenAI models?

The difference between retirement and deprecation of a model is important to understand:

  • Retirement: This refers to when a model is completely removed and is no longer available for use. After the retirement date, clients can no longer use the model as it will be removed from the Azure service.

  • Deprecation: Deprecation occurs before retirement. After a model is deprecated, it can no longer be deployed or provisioned for new clients. However, if a client is already using the model, they can continue to do so until the retirement date. No new deployments of the deprecated model are allowed, but existing deployments will still function until the model is fully retired.

Can models be customized/bring our own models?

Yes, we even have customers that are doing this.

Is there a platform available for conducting automated tests and comparing results for custom models?

Yes, through benchmarking, it is possible to perform comparisons.

There is a documentation for this process: https://unique-ch.atlassian.net/wiki/x/AQDJIw

How can one train, test, and deploy a model for use with Unique’s solution?

So far, we have not directly trained a model ourselves; instead, our customers have undertaken this task. However, our Data-Science team has provided support and guidance to them throughout the process.

Does one need to use Azure AI Studio?

There's no need to restrict yourself to Azure AI Studio exclusively. As long as the model can be provisioned, we are capable of integrating it.

Is it possible to implement version control for models, such as maintaining a development version, publishing a beta version, and continuing to use a previous version?

Yes, within the system, each prompt allows for the selection of the model and its version at will. We practice this on a regular basis, especially with the release of new minor or preview versions from Azure OpenAI.

Can models be shared by user-groups?

Yes, it can be scoped by user-groups.

Can models be restricted by
user-groups?

Yes, it can be restricted by user-groups.

Can token consumption be followed by model or by user-groups?

This feature is currently under development and not available yet.

However, an Analytics Framework with downloadable CSV-Reports is already in place and covers these points:

  • User Engagement

  • Assistant Usage

  • Most referenced files

Read more about this here: Analytics

A report incl. consumption by assistant/model is planned for Q2 2024.

Are there several types of prices depending on the models used?

Our pricing model remains fixed, however, the costs of the underlying models set by Microsoft are subject to change and are transparently communicated back to you. Prices may fluctuate. We offer guidance on which prompts require specific models.

How is visibility kept on the costs related to the API usage?

We report the costs generated on the Subscription on a monthly basis. In the early days of the project, we can negotiate a faster rhythm.

Is it possible to set token limits for each model or user group, including actions like sending alerts or shutting down the API?

This feature is not available yet and is currently under development, planned for Q3 2024.

Is it possible to grant standard access to ChatGPT-3.5, replacing the direct access currently provided to certain staff members?

Yes, this is even included in the base configuration of Unique. You can even give access to ChatGPT-4.

Is your solution offered on the MS Marketplace?

Unique is currently not offered in the MS Marketplace.

What tests have been done to select the appropriate LLM models?

We conducted benchmarks using our documents, and our clients performed similar tests. This process helps us select the most suitable models for each prompt or use case. While we have evaluated other models, we found that they do not yet match the performance of GPT-4, especially in situations requiring RAG.

...