mirror of
https://github.com/microsoft/PowerToys.git
synced 2025-12-25 16:09:54 +01:00
<!-- Enter a brief description/summary of your PR here. What does it fix/what does it change/how was it tested (even manually, if necessary)? --> ## Summary of the Pull Request Cache of the downloaded model will make the newly added model only work after running of powertoys, this disable the cache, so just downloaded model will take effect immediately <!-- Please review the items on the PR checklist before submitting--> ## PR Checklist - [ ] Closes: #xxx <!-- - [ ] Closes: #yyy (add separate lines for additional resolved issues) --> - [ ] **Communication:** I've discussed this with core contributors already. If the work hasn't been agreed, this work might be rejected - [ ] **Tests:** Added/updated and all pass - [ ] **Localization:** All end-user-facing strings can be localized - [ ] **Dev docs:** Added/updated - [ ] **New binaries:** Added on the required places - [ ] [JSON for signing](https://github.com/microsoft/PowerToys/blob/main/.pipelines/ESRPSigning_core.json) for new binaries - [ ] [WXS for installer](https://github.com/microsoft/PowerToys/blob/main/installer/PowerToysSetup/Product.wxs) for new binaries and localization folder - [ ] [YML for CI pipeline](https://github.com/microsoft/PowerToys/blob/main/.pipelines/ci/templates/build-powertoys-steps.yml) for new test projects - [ ] [YML for signed pipeline](https://github.com/microsoft/PowerToys/blob/main/.pipelines/release.yml) - [ ] **Documentation updated:** If checked, please file a pull request on [our docs repo](https://github.com/MicrosoftDocs/windows-uwp/tree/docs/hub/powertoys) and link it here: #xxx <!-- Provide a more detailed description of the PR, other things fixed, or any additional comments/features here --> ## Detailed Description of the Pull Request / Additional comments <!-- Describe how you validated the behavior. Add automated tests wherever possible, but list manual validation steps taken as well --> ## Validation Steps Performed Validated locally
157 lines
5.6 KiB
C#
157 lines
5.6 KiB
C#
// Copyright (c) Microsoft Corporation
|
|
// The Microsoft Corporation licenses this file to you under the MIT license.
|
|
// See the LICENSE file in the project root for more information.
|
|
|
|
using System.ClientModel;
|
|
using LanguageModelProvider.FoundryLocal;
|
|
using ManagedCommon;
|
|
using Microsoft.Extensions.AI;
|
|
using OpenAI;
|
|
|
|
namespace LanguageModelProvider;
|
|
|
|
public sealed class FoundryLocalModelProvider : ILanguageModelProvider
|
|
{
|
|
private FoundryClient? _foundryClient;
|
|
private IEnumerable<FoundryCatalogModel>? _catalogModels;
|
|
private string? _serviceUrl;
|
|
|
|
public static FoundryLocalModelProvider Instance { get; } = new();
|
|
|
|
public string Name => "FoundryLocal";
|
|
|
|
public string ProviderDescription => "The model will run locally via Foundry Local";
|
|
|
|
public IChatClient? GetIChatClient(string modelId)
|
|
{
|
|
Logger.LogInfo($"[FoundryLocal] GetIChatClient called with url: {modelId}");
|
|
InitializeAsync().GetAwaiter().GetResult();
|
|
|
|
if (string.IsNullOrWhiteSpace(modelId))
|
|
{
|
|
Logger.LogError("[FoundryLocal] Model ID is empty after extraction");
|
|
return null;
|
|
}
|
|
|
|
// Check if model is in catalog
|
|
var isInCatalog = _catalogModels?.Any(m => m.Name == modelId) ?? false;
|
|
if (!isInCatalog)
|
|
{
|
|
var errorMessage = $"{modelId} is not supported in Foundry Local. Please configure supported models in Settings.";
|
|
Logger.LogError($"[FoundryLocal] {errorMessage}");
|
|
throw new InvalidOperationException(errorMessage);
|
|
}
|
|
|
|
// Ensure the model is loaded before returning chat client
|
|
var isLoaded = _foundryClient!.EnsureModelLoaded(modelId).GetAwaiter().GetResult();
|
|
if (!isLoaded)
|
|
{
|
|
Logger.LogError($"[FoundryLocal] Failed to load model: {modelId}");
|
|
throw new InvalidOperationException($"Failed to load the model '{modelId}'.");
|
|
}
|
|
|
|
// Use ServiceUri instead of Endpoint since Endpoint already includes /v1
|
|
var baseUri = _foundryClient.GetServiceUri();
|
|
if (baseUri == null)
|
|
{
|
|
const string message = "Foundry Local service URL is not available. Please make sure Foundry Local is installed and running.";
|
|
Logger.LogError($"[FoundryLocal] {message}");
|
|
throw new InvalidOperationException(message);
|
|
}
|
|
|
|
var endpointUri = new Uri($"{baseUri.ToString().TrimEnd('/')}/v1");
|
|
Logger.LogInfo($"[FoundryLocal] Creating OpenAI client with endpoint: {endpointUri}");
|
|
|
|
return new OpenAIClient(
|
|
new ApiKeyCredential("none"),
|
|
new OpenAIClientOptions { Endpoint = endpointUri })
|
|
.GetChatClient(modelId)
|
|
.AsIChatClient();
|
|
}
|
|
|
|
public string GetIChatClientString(string url)
|
|
{
|
|
try
|
|
{
|
|
InitializeAsync().GetAwaiter().GetResult();
|
|
}
|
|
catch
|
|
{
|
|
return string.Empty;
|
|
}
|
|
|
|
var modelId = url.Split('/').LastOrDefault();
|
|
|
|
if (string.IsNullOrWhiteSpace(_serviceUrl) || string.IsNullOrWhiteSpace(modelId))
|
|
{
|
|
return string.Empty;
|
|
}
|
|
|
|
return $"new OpenAIClient(new ApiKeyCredential(\"none\"), new OpenAIClientOptions{{ Endpoint = new Uri(\"{_serviceUrl}/v1\") }}).GetChatClient(\"{modelId}\").AsIChatClient()";
|
|
}
|
|
|
|
public async Task<IEnumerable<ModelDetails>> GetModelsAsync(CancellationToken cancelationToken = default)
|
|
{
|
|
await InitializeAsync(cancelationToken);
|
|
|
|
if (_foundryClient == null)
|
|
{
|
|
return Array.Empty<ModelDetails>();
|
|
}
|
|
|
|
var cachedModels = await _foundryClient.ListCachedModels();
|
|
List<ModelDetails> downloadedModels = [];
|
|
|
|
foreach (var model in cachedModels)
|
|
{
|
|
Logger.LogInfo($"[FoundryLocal] Adding unmatched cached model: {model.Name}");
|
|
downloadedModels.Add(new ModelDetails
|
|
{
|
|
Id = $"fl-{model.Name}",
|
|
Name = model.Name,
|
|
Url = $"fl://{model.Name}",
|
|
Description = $"{model.Name} running locally with Foundry Local",
|
|
HardwareAccelerators = [HardwareAccelerator.FOUNDRYLOCAL],
|
|
ProviderModelDetails = model,
|
|
});
|
|
}
|
|
|
|
return downloadedModels;
|
|
}
|
|
|
|
private async Task InitializeAsync(CancellationToken cancelationToken = default)
|
|
{
|
|
if (_foundryClient != null && _catalogModels != null && _catalogModels.Any())
|
|
{
|
|
await _foundryClient.EnsureRunning().ConfigureAwait(false);
|
|
return;
|
|
}
|
|
|
|
Logger.LogInfo("[FoundryLocal] Initializing provider");
|
|
_foundryClient ??= await FoundryClient.CreateAsync();
|
|
|
|
if (_foundryClient == null)
|
|
{
|
|
const string message = "Foundry Local client could not be created. Please make sure Foundry Local is installed and running.";
|
|
Logger.LogError($"[FoundryLocal] {message}");
|
|
throw new InvalidOperationException(message);
|
|
}
|
|
|
|
_serviceUrl ??= await _foundryClient.GetServiceUrl();
|
|
Logger.LogInfo($"[FoundryLocal] Service URL: {_serviceUrl}");
|
|
|
|
var catalogModels = await _foundryClient.ListCatalogModels();
|
|
Logger.LogInfo($"[FoundryLocal] Found {catalogModels.Count} catalog models");
|
|
_catalogModels = catalogModels;
|
|
}
|
|
|
|
public async Task<bool> IsAvailable()
|
|
{
|
|
Logger.LogInfo("[FoundryLocal] Checking availability");
|
|
await InitializeAsync();
|
|
var available = _foundryClient != null;
|
|
Logger.LogInfo($"[FoundryLocal] Available: {available}");
|
|
return available;
|
|
}
|
|
}
|