diff options
| author | Aldrik Ramaekers <aldrik@mailbox.org> | 2026-01-09 11:45:12 +0100 |
|---|---|---|
| committer | Aldrik Ramaekers <aldrik@mailbox.org> | 2026-01-09 11:45:12 +0100 |
| commit | a65e876f3277a7d7fca6e5129ac3e200dae2d0dc (patch) | |
| tree | 3d777dca9e07bf1fbd4a8641e1a4719094ac7789 /src/providers | |
| parent | 8bdab613289e2626173de8bff31e7078dd03a16e (diff) | |
save all ai service configs
Diffstat (limited to 'src/providers')
| -rw-r--r-- | src/providers/DeepSeek.cpp | 4 | ||||
| -rw-r--r-- | src/providers/openAI.cpp | 30 |
2 files changed, 26 insertions, 8 deletions
diff --git a/src/providers/DeepSeek.cpp b/src/providers/DeepSeek.cpp index c34e299..8a5b42e 100644 --- a/src/providers/DeepSeek.cpp +++ b/src/providers/DeepSeek.cpp @@ -30,7 +30,7 @@ static bool _DeepSeek_query_with_file(const char* query, size_t query_length, ch (void)query_length; assert(query_buffer); - const char *api_key = administration::get_ai_service().api_key_public; + const char *api_key = administration::get_active_ai_service().api_key_public; httplib::SSLClient cli("api.deepseek.com"); //cli.enable_server_certificate_verification(false); @@ -46,7 +46,7 @@ static bool _DeepSeek_query_with_file(const char* query, size_t query_length, ch size_t body_size = file_size + QUERY_BUFFER_SIZE; char* body = (char*)memops::alloc(body_size); strops::format(body, body_size, - "{\"model\":\"%s\", \"messages\": [ { \"role\": \"user\", \"content\": \"%s\" } ] }", administration::get_ai_service().model_name, query_escaped); + "{\"model\":\"%s\", \"messages\": [ { \"role\": \"user\", \"content\": \"%s\" } ] }", administration::get_active_ai_service().model_name, query_escaped); httplib::Headers headers; headers.insert(std::make_pair("Authorization", std::string("Bearer ") + api_key)); diff --git a/src/providers/openAI.cpp b/src/providers/openAI.cpp index d1495dc..6c60541 100644 --- a/src/providers/openAI.cpp +++ b/src/providers/openAI.cpp @@ -26,7 +26,7 @@ static bool _openAI_batch_query_with_file(const char** queries, size_t query_count, char* file_id, invoice* buffer, importer::batch_query_response_handler response_handler) { - const char *api_key = administration::get_ai_service().api_key_public; + const char *api_key = administration::get_active_ai_service().api_key_public; httplib::SSLClient cli("api.openai.com", 443); thrd_t threads[20]; @@ -50,7 +50,7 @@ static bool _openAI_batch_query_with_file(const char** queries, size_t query_cou " }" "], " " \"text\": { \"format\": { \"type\": \"json_object\" } } " - "}", administration::get_ai_service().model_name, file_id, query_escaped); + "}", administration::get_active_ai_service().model_name, file_id, query_escaped); httplib::Headers headers; headers.insert(std::make_pair("Authorization", std::string("Bearer ") + api_key)); @@ -96,7 +96,7 @@ static bool _openAI_batch_query_with_file(const char** queries, size_t query_cou static bool _openAI_query_with_file(const char* query, size_t query_length, char* file_id, char** response) { - const char *api_key = administration::get_ai_service().api_key_public; + const char *api_key = administration::get_active_ai_service().api_key_public; httplib::SSLClient cli("api.openai.com", 443); //cli.enable_server_certificate_verification(false); @@ -108,7 +108,7 @@ static bool _openAI_query_with_file(const char* query, size_t query_length, char char* body = (char*)memops::alloc(body_size); strops::format(body, body_size, "{\"model\":\"%s\", \"input\": [ { \"role\": \"user\", \"content\": [ { \"type\": \"input_file\", \"file_id\": \"%s\" }, " - "{ \"type\": \"input_text\", \"text\": \"%s\" } ] } ] }", administration::get_ai_service().model_name, file_id, query_escaped); + "{ \"type\": \"input_text\", \"text\": \"%s\" } ] } ] }", administration::get_active_ai_service().model_name, file_id, query_escaped); httplib::Headers headers; headers.insert(std::make_pair("Authorization", std::string("Bearer ") + api_key)); @@ -135,7 +135,7 @@ static bool _openAI_query_with_file(const char* query, size_t query_length, char static bool _openAI_upload_file(char* file_path, char* file_id, size_t file_id_len) { - const char *api_key = administration::get_ai_service().api_key_public; + const char *api_key = administration::get_active_ai_service().api_key_public; const char *filename = strops::get_filename(file_path); FILE* orig_file = fopen(file_path, "rb"); @@ -237,7 +237,7 @@ static bool _openAI_upload_file(char* file_path, char* file_id, size_t file_id_l static bool _openAI_get_available_models(importer::model_list_request* buffer) { - const char *api_key = administration::get_ai_service().api_key_public; + const char *api_key = administration::get_active_ai_service().api_key_public; httplib::SSLClient cli("api.openai.com", 443); @@ -273,4 +273,22 @@ importer::ai_provider_impl _chatgpt_api_provider = { _openAI_query_with_file, _openAI_batch_query_with_file, _openAI_get_available_models, +}; + +importer::ai_provider_impl _gemini_api_provider = { + "Gemini", + "", + 0, + 0, + 0, + 0, +}; + +importer::ai_provider_impl _perplexity_api_provider = { + "Perplexity", + "", + 0, + 0, + 0, + 0, };
\ No newline at end of file |
