11#include "absl/strings/str_cat.h"
12#include "absl/strings/match.h"
13#include "absl/strings/str_format.h"
14#include "absl/strings/str_split.h"
15#include "absl/strings/strip.h"
16#include "absl/time/clock.h"
17#include "absl/time/time.h"
22#include <TargetConditionals.h>
25#if defined(__APPLE__) && (TARGET_OS_IPHONE == 1 || TARGET_IPHONE_SIMULATOR == 1)
27#define YAZE_AI_IOS_URLSESSION 1
35#include "nlohmann/json.hpp"
38#ifdef CPPHTTPLIB_OPENSSL_SUPPORT
39#include <openssl/crypto.h>
40#include <openssl/err.h>
41#include <openssl/ssl.h>
44static std::atomic<bool> g_openssl_initialized{
false};
45static std::mutex g_openssl_init_mutex;
47static void EnsureOpenSSLInitialized() {
48 std::lock_guard<std::mutex> lock(g_openssl_init_mutex);
49 if (!g_openssl_initialized.exchange(
true)) {
51 OPENSSL_INIT_LOAD_SSL_STRINGS | OPENSSL_INIT_LOAD_CRYPTO_STRINGS,
53 std::cerr <<
"✓ OpenSSL initialized for HTTPS support" << std::endl;
62#ifdef YAZE_AI_RUNTIME_AVAILABLE
65 : function_calling_enabled_(config.use_function_calling), config_(config) {
66 if (config_.verbose) {
67 std::cerr <<
"[DEBUG] Initializing OpenAI service..." << std::endl;
68 std::cerr <<
"[DEBUG] Model: " << config_.model << std::endl;
69 std::cerr <<
"[DEBUG] Function calling: "
70 << (function_calling_enabled_ ?
"enabled" :
"disabled")
74#ifdef CPPHTTPLIB_OPENSSL_SUPPORT
75 EnsureOpenSSLInitialized();
76 if (config_.verbose) {
77 std::cerr <<
"[DEBUG] OpenSSL initialized for HTTPS" << std::endl;
82 std::string catalogue_path = config_.prompt_version ==
"v2"
83 ?
"assets/agent/prompt_catalogue_v2.yaml"
84 :
"assets/agent/prompt_catalogue.yaml";
85 if (
auto status = prompt_builder_.LoadResourceCatalogue(catalogue_path);
87 std::cerr <<
"⚠️ Failed to load agent prompt catalogue: "
88 << status.message() << std::endl;
91 if (config_.system_instruction.empty()) {
93 std::string prompt_file;
94 if (config_.prompt_version ==
"v3") {
95 prompt_file =
"agent/system_prompt_v3.txt";
96 }
else if (config_.prompt_version ==
"v2") {
97 prompt_file =
"agent/system_prompt_v2.txt";
99 prompt_file =
"agent/system_prompt.txt";
102 auto prompt_path = util::PlatformPaths::FindAsset(prompt_file);
103 if (prompt_path.ok()) {
104 std::ifstream file(prompt_path->string());
106 std::stringstream buffer;
107 buffer << file.rdbuf();
108 config_.system_instruction = buffer.str();
109 if (config_.verbose) {
110 std::cerr <<
"[DEBUG] Loaded prompt: " << prompt_path->string()
116 if (config_.system_instruction.empty()) {
117 config_.system_instruction = BuildSystemInstruction();
121 if (config_.verbose) {
122 std::cerr <<
"[DEBUG] OpenAI service initialized" << std::endl;
126void OpenAIAIService::EnableFunctionCalling(
bool enable) {
127 function_calling_enabled_ = enable;
130std::vector<std::string> OpenAIAIService::GetAvailableTools()
const {
131 return {
"resource-list",
"resource-search",
132 "dungeon-list-sprites",
"dungeon-describe-room",
133 "overworld-find-tile",
"overworld-describe-map",
134 "overworld-list-warps"};
137std::string OpenAIAIService::BuildFunctionCallSchemas() {
138#ifndef YAZE_WITH_JSON
141 std::string schemas = prompt_builder_.BuildFunctionCallSchemas();
142 if (!schemas.empty() && schemas !=
"[]") {
146 auto schema_path_or =
147 util::PlatformPaths::FindAsset(
"agent/function_schemas.json");
149 if (!schema_path_or.ok()) {
153 std::ifstream file(schema_path_or->string());
154 if (!file.is_open()) {
159 nlohmann::json schemas_json;
160 file >> schemas_json;
161 return schemas_json.dump();
162 }
catch (
const nlohmann::json::exception& e) {
163 std::cerr <<
"⚠️ Failed to parse function schemas JSON: " << e.what()
170std::string OpenAIAIService::BuildSystemInstruction() {
171 return prompt_builder_.BuildSystemInstruction();
174void OpenAIAIService::SetRomContext(Rom* rom) {
175 prompt_builder_.SetRom(rom);
178absl::StatusOr<std::vector<ModelInfo>> OpenAIAIService::ListAvailableModels() {
179#ifndef YAZE_WITH_JSON
180 return absl::UnimplementedError(
"OpenAI AI service requires JSON support");
182 const bool is_openai_cloud =
183 absl::StrContains(config_.base_url,
"api.openai.com");
184 if (config_.api_key.empty() && is_openai_cloud) {
186 std::vector<ModelInfo> defaults = {
188 .display_name =
"GPT-4o",
189 .provider =
"openai",
190 .description =
"Most capable GPT-4 model"},
191 {.name =
"gpt-4o-mini",
192 .display_name =
"GPT-4o Mini",
193 .provider =
"openai",
194 .description =
"Fast and cost-effective"},
195 {.name =
"gpt-4-turbo",
196 .display_name =
"GPT-4 Turbo",
197 .provider =
"openai",
198 .description =
"GPT-4 with larger context"},
199 {.name =
"gpt-3.5-turbo",
200 .display_name =
"GPT-3.5 Turbo",
201 .provider =
"openai",
202 .description =
"Fast and efficient"}};
207 if (config_.verbose) {
208 std::cerr <<
"[DEBUG] Listing OpenAI models..." << std::endl;
211 std::string response_str;
212#if defined(YAZE_AI_IOS_URLSESSION)
213 std::map<std::string, std::string> headers;
214 if (!config_.api_key.empty()) {
215 headers.emplace(
"Authorization",
"Bearer " + config_.api_key);
217 auto resp_or = ios::UrlSessionHttpRequest(
218 "GET", config_.base_url +
"/v1/models", headers,
"", 8000);
220 if (config_.verbose) {
221 std::cerr <<
"[DEBUG] OpenAI /v1/models failed: "
222 << resp_or.status().message() << std::endl;
225 std::vector<ModelInfo> defaults = {
226 {.name =
"gpt-4o-mini",
227 .display_name =
"GPT-4o Mini",
228 .provider =
"openai"},
229 {.name =
"gpt-4o", .display_name =
"GPT-4o", .provider =
"openai"},
230 {.name =
"gpt-3.5-turbo",
231 .display_name =
"GPT-3.5 Turbo",
232 .provider =
"openai"}};
235 if (resp_or->status_code != 200) {
236 if (config_.verbose) {
237 std::cerr <<
"[DEBUG] OpenAI /v1/models HTTP " << resp_or->status_code
240 std::vector<ModelInfo> defaults = {
241 {.name =
"gpt-4o-mini",
242 .display_name =
"GPT-4o Mini",
243 .provider =
"openai"},
244 {.name =
"gpt-4o", .display_name =
"GPT-4o", .provider =
"openai"},
245 {.name =
"gpt-3.5-turbo",
246 .display_name =
"GPT-3.5 Turbo",
247 .provider =
"openai"}};
250 response_str = resp_or->body;
253 std::string auth_header = config_.api_key.empty()
255 :
"-H 'Authorization: Bearer " + config_.api_key +
"' ";
256 std::string curl_cmd =
257 "curl -s -X GET '" + config_.base_url +
"/v1/models' " +
258 auth_header +
"2>&1";
261 FILE* pipe = _popen(curl_cmd.c_str(),
"r");
263 FILE* pipe = popen(curl_cmd.c_str(),
"r");
266 return absl::InternalError(
"Failed to execute curl command");
270 while (fgets(buffer,
sizeof(buffer), pipe) !=
nullptr) {
271 response_str += buffer;
281 auto models_json = nlohmann::json::parse(response_str,
nullptr,
false);
282 if (models_json.is_discarded()) {
283 return absl::InternalError(
"Failed to parse OpenAI models JSON");
286 if (!models_json.contains(
"data")) {
288 std::vector<ModelInfo> defaults = {
289 {.name =
"gpt-4o-mini",
290 .display_name =
"GPT-4o Mini",
291 .provider =
"openai"},
292 {.name =
"gpt-4o", .display_name =
"GPT-4o", .provider =
"openai"},
293 {.name =
"gpt-3.5-turbo",
294 .display_name =
"GPT-3.5 Turbo",
295 .provider =
"openai"}};
299 std::vector<ModelInfo> models;
300 for (
const auto& m : models_json[
"data"]) {
301 std::string
id = m.value(
"id",
"");
305 bool is_local = !absl::StrContains(config_.base_url,
"api.openai.com");
307 if (is_local || absl::StartsWith(
id,
"gpt-4") || absl::StartsWith(
id,
"gpt-3.5") ||
308 absl::StartsWith(
id,
"o1") || absl::StartsWith(
id,
"chatgpt")) {
311 info.display_name = id;
312 info.provider =
"openai";
313 info.family = is_local ?
"local" :
"gpt";
314 info.is_local = is_local;
318 info.display_name =
"GPT-4o";
319 else if (
id ==
"gpt-4o-mini")
320 info.display_name =
"GPT-4o Mini";
321 else if (
id ==
"gpt-4-turbo")
322 info.display_name =
"GPT-4 Turbo";
323 else if (
id ==
"gpt-3.5-turbo")
324 info.display_name =
"GPT-3.5 Turbo";
325 else if (
id ==
"o1-preview")
326 info.display_name =
"o1 Preview";
327 else if (
id ==
"o1-mini")
328 info.display_name =
"o1 Mini";
330 models.push_back(std::move(info));
335 }
catch (
const std::exception& e) {
336 return absl::InternalError(
337 absl::StrCat(
"Failed to list models: ", e.what()));
342absl::Status OpenAIAIService::CheckAvailability() {
343#ifndef YAZE_WITH_JSON
344 return absl::UnimplementedError(
345 "OpenAI AI service requires JSON support. Build with "
346 "-DYAZE_WITH_JSON=ON");
350 bool is_local_server = config_.base_url !=
"https://api.openai.com";
351 if (config_.api_key.empty() && !is_local_server) {
352 return absl::FailedPreconditionError(
353 "❌ OpenAI API key not configured\n"
354 " Set OPENAI_API_KEY environment variable\n"
355 " Get your API key at: https://platform.openai.com/api-keys\n"
356 " For LMStudio, use --openai_base_url=http://localhost:1234");
360#if defined(YAZE_AI_IOS_URLSESSION)
361 std::map<std::string, std::string> headers;
362 if (!config_.api_key.empty()) {
363 headers.emplace(
"Authorization",
"Bearer " + config_.api_key);
365 auto resp_or = ios::UrlSessionHttpRequest(
366 "GET", config_.base_url +
"/v1/models", headers,
"", 8000);
368 return absl::UnavailableError(
369 absl::StrCat(
"❌ Cannot reach OpenAI API\n ",
370 resp_or.status().message()));
372 if (resp_or->status_code == 401) {
373 return absl::PermissionDeniedError(
374 "❌ Invalid OpenAI API key\n"
375 " Verify your key at: https://platform.openai.com/api-keys");
377 if (resp_or->status_code != 200) {
378 return absl::InternalError(absl::StrCat(
379 "❌ OpenAI API error: ", resp_or->status_code,
"\n ",
383 httplib::Client cli(config_.base_url);
384 cli.set_connection_timeout(5, 0);
386 httplib::Headers headers = {};
387 if (!config_.api_key.empty()) {
388 headers.emplace(
"Authorization",
"Bearer " + config_.api_key);
391 auto res = cli.Get(
"/v1/models", headers);
394 return absl::UnavailableError(
395 "❌ Cannot reach OpenAI API\n"
396 " Check your internet connection");
399 if (res->status == 401) {
400 return absl::PermissionDeniedError(
401 "❌ Invalid OpenAI API key\n"
402 " Verify your key at: https://platform.openai.com/api-keys");
405 if (res->status != 200) {
406 return absl::InternalError(absl::StrCat(
407 "❌ OpenAI API error: ", res->status,
"\n ", res->body));
411 return absl::OkStatus();
412 }
catch (
const std::exception& e) {
413 return absl::InternalError(
414 absl::StrCat(
"Exception during availability check: ", e.what()));
419absl::StatusOr<AgentResponse> OpenAIAIService::GenerateResponse(
420 const std::string& prompt) {
421 return GenerateResponse(
422 {{{agent::ChatMessage::Sender::kUser, prompt, absl::Now()}}});
425absl::StatusOr<AgentResponse> OpenAIAIService::GenerateResponse(
426 const std::vector<agent::ChatMessage>& history) {
427#ifndef YAZE_WITH_JSON
428 return absl::UnimplementedError(
429 "OpenAI AI service requires JSON support. Build with "
430 "-DYAZE_WITH_JSON=ON");
432 if (history.empty()) {
433 return absl::InvalidArgumentError(
"History cannot be empty.");
436 const bool is_openai_cloud =
437 absl::StrContains(config_.base_url,
"api.openai.com");
438 if (config_.api_key.empty() && is_openai_cloud) {
439 return absl::FailedPreconditionError(
"OpenAI API key not configured");
442 absl::Time request_start = absl::Now();
445 if (config_.verbose) {
446 std::cerr <<
"[DEBUG] Using curl for OpenAI HTTPS request" << std::endl;
447 std::cerr <<
"[DEBUG] Processing " << history.size()
448 <<
" messages in history" << std::endl;
452 nlohmann::json messages = nlohmann::json::array();
456 {{
"role",
"system"}, {
"content", config_.system_instruction}});
459 int start_idx = std::max(0,
static_cast<int>(history.size()) - 10);
460 for (
size_t i = start_idx; i < history.size(); ++i) {
461 const auto& msg = history[i];
462 std::string role = (msg.sender == agent::ChatMessage::Sender::kUser)
466 messages.push_back({{
"role", role}, {
"content", msg.message}});
470 nlohmann::json request_body = {{
"model", config_.model},
471 {
"messages", messages},
472 {
"temperature", config_.temperature},
473 {
"max_tokens", config_.max_output_tokens}};
476 if (function_calling_enabled_) {
478 std::string schemas_str = BuildFunctionCallSchemas();
479 if (config_.verbose) {
480 std::cerr <<
"[DEBUG] Function calling schemas: "
481 << schemas_str.substr(0, 200) <<
"..." << std::endl;
484 nlohmann::json schemas = nlohmann::json::parse(schemas_str);
486 if (schemas.is_array() && !schemas.empty()) {
488 nlohmann::json tools = nlohmann::json::array();
489 for (
const auto& schema : schemas) {
490 tools.push_back({{
"type",
"function"}, {
"function", schema}});
492 request_body[
"tools"] = tools;
494 }
catch (
const nlohmann::json::exception& e) {
495 std::cerr <<
"⚠️ Failed to parse function schemas: " << e.what()
500 if (config_.verbose) {
501 std::cerr <<
"[DEBUG] Sending " << messages.size()
502 <<
" messages to OpenAI" << std::endl;
505 std::string response_str;
506#if defined(YAZE_AI_IOS_URLSESSION)
507 std::map<std::string, std::string> headers;
508 headers.emplace(
"Content-Type",
"application/json");
509 if (!config_.api_key.empty()) {
510 headers.emplace(
"Authorization",
"Bearer " + config_.api_key);
512 auto resp_or = ios::UrlSessionHttpRequest(
513 "POST", config_.base_url +
"/v1/chat/completions", headers,
514 request_body.dump(), 60000);
516 return resp_or.status();
518 if (resp_or->status_code == 401) {
519 return absl::PermissionDeniedError(
520 "❌ Invalid OpenAI API key\n"
521 " Verify your key at: https://platform.openai.com/api-keys");
523 if (resp_or->status_code != 200) {
524 return absl::InternalError(absl::StrCat(
525 "❌ OpenAI API error: ", resp_or->status_code,
"\n ",
528 response_str = resp_or->body;
531 std::string temp_file =
"/tmp/openai_request.json";
532 std::ofstream out(temp_file);
533 out << request_body.dump();
537 std::string auth_header = config_.api_key.empty()
539 :
"-H 'Authorization: Bearer " + config_.api_key +
"' ";
540 std::string curl_cmd =
541 "curl -s -X POST '" + config_.base_url +
"/v1/chat/completions' "
542 "-H 'Content-Type: application/json' " +
547 if (config_.verbose) {
548 std::cerr <<
"[DEBUG] Executing OpenAI API request..." << std::endl;
552 FILE* pipe = _popen(curl_cmd.c_str(),
"r");
554 FILE* pipe = popen(curl_cmd.c_str(),
"r");
557 return absl::InternalError(
"Failed to execute curl command");
561 while (fgets(buffer,
sizeof(buffer), pipe) !=
nullptr) {
562 response_str += buffer;
566 int status = _pclose(pipe);
568 int status = pclose(pipe);
570 std::remove(temp_file.c_str());
573 return absl::InternalError(
574 absl::StrCat(
"Curl failed with status ", status));
578 if (response_str.empty()) {
579 return absl::InternalError(
"Empty response from OpenAI API");
582 if (config_.verbose) {
585 <<
"🔍 Raw OpenAI API Response:"
588 <<
"\033[2m" << response_str.substr(0, 500) <<
"\033[0m"
592 if (config_.verbose) {
593 std::cerr <<
"[DEBUG] Parsing response..." << std::endl;
596 auto parsed_or = ParseOpenAIResponse(response_str);
597 if (!parsed_or.ok()) {
598 return parsed_or.status();
601 AgentResponse agent_response = std::move(parsed_or.value());
602 agent_response.provider =
"openai";
603 agent_response.model = config_.model;
604 agent_response.latency_seconds =
605 absl::ToDoubleSeconds(absl::Now() - request_start);
606 agent_response.parameters[
"prompt_version"] = config_.prompt_version;
607 agent_response.parameters[
"temperature"] =
608 absl::StrFormat(
"%.2f", config_.temperature);
609 agent_response.parameters[
"max_output_tokens"] =
610 absl::StrFormat(
"%d", config_.max_output_tokens);
611 agent_response.parameters[
"function_calling"] =
612 function_calling_enabled_ ?
"true" :
"false";
614 return agent_response;
616 }
catch (
const std::exception& e) {
617 if (config_.verbose) {
618 std::cerr <<
"[ERROR] Exception: " << e.what() << std::endl;
620 return absl::InternalError(
621 absl::StrCat(
"Exception during generation: ", e.what()));
626absl::StatusOr<AgentResponse> OpenAIAIService::ParseOpenAIResponse(
627 const std::string& response_body) {
628#ifndef YAZE_WITH_JSON
629 return absl::UnimplementedError(
"JSON support required");
631 AgentResponse agent_response;
633 auto response_json = nlohmann::json::parse(response_body,
nullptr,
false);
634 if (response_json.is_discarded()) {
635 return absl::InternalError(
"❌ Failed to parse OpenAI response JSON");
639 if (response_json.contains(
"error")) {
640 std::string error_msg =
641 response_json[
"error"].value(
"message",
"Unknown error");
642 return absl::InternalError(absl::StrCat(
"❌ OpenAI API error: ", error_msg));
646 if (!response_json.contains(
"choices") || response_json[
"choices"].empty()) {
647 return absl::InternalError(
"❌ No choices in OpenAI response");
650 const auto& choice = response_json[
"choices"][0];
651 if (!choice.contains(
"message")) {
652 return absl::InternalError(
"❌ No message in OpenAI response");
655 const auto& message = choice[
"message"];
658 if (message.contains(
"content") && !message[
"content"].is_null()) {
659 std::string text_content = message[
"content"].get<std::string>();
661 if (config_.verbose) {
664 <<
"🔍 Raw LLM Response:"
667 <<
"\033[2m" << text_content <<
"\033[0m"
672 text_content = std::string(absl::StripAsciiWhitespace(text_content));
673 if (absl::StartsWith(text_content,
"```json")) {
674 text_content = text_content.substr(7);
675 }
else if (absl::StartsWith(text_content,
"```")) {
676 text_content = text_content.substr(3);
678 if (absl::EndsWith(text_content,
"```")) {
679 text_content = text_content.substr(0, text_content.length() - 3);
681 text_content = std::string(absl::StripAsciiWhitespace(text_content));
684 auto parsed_text = nlohmann::json::parse(text_content,
nullptr,
false);
685 if (!parsed_text.is_discarded()) {
687 if (parsed_text.contains(
"text_response") &&
688 parsed_text[
"text_response"].is_string()) {
689 agent_response.text_response =
690 parsed_text[
"text_response"].get<std::string>();
694 if (parsed_text.contains(
"reasoning") &&
695 parsed_text[
"reasoning"].is_string()) {
696 agent_response.reasoning = parsed_text[
"reasoning"].get<std::string>();
700 if (parsed_text.contains(
"commands") &&
701 parsed_text[
"commands"].is_array()) {
702 for (
const auto& cmd : parsed_text[
"commands"]) {
703 if (cmd.is_string()) {
704 std::string command = cmd.get<std::string>();
705 if (absl::StartsWith(command,
"z3ed ")) {
706 command = command.substr(5);
708 agent_response.commands.push_back(command);
714 if (parsed_text.contains(
"tool_calls") &&
715 parsed_text[
"tool_calls"].is_array()) {
716 for (
const auto& call : parsed_text[
"tool_calls"]) {
717 if (call.contains(
"tool_name") && call[
"tool_name"].is_string()) {
719 tool_call.tool_name = call[
"tool_name"].get<std::string>();
720 if (call.contains(
"args") && call[
"args"].is_object()) {
721 for (
auto& [key, value] : call[
"args"].items()) {
722 if (value.is_string()) {
723 tool_call.args[
key] = value.get<std::string>();
724 }
else if (value.is_number()) {
725 tool_call.args[
key] = std::to_string(value.get<
double>());
726 }
else if (value.is_boolean()) {
727 tool_call.args[
key] = value.get<
bool>() ?
"true" :
"false";
731 agent_response.tool_calls.push_back(tool_call);
737 agent_response.text_response = text_content;
742 if (message.contains(
"tool_calls") && message[
"tool_calls"].is_array()) {
743 for (
const auto& call : message[
"tool_calls"]) {
744 if (call.contains(
"function")) {
745 const auto& func = call[
"function"];
747 tool_call.tool_name = func.value(
"name",
"");
749 if (func.contains(
"arguments") && func[
"arguments"].is_string()) {
750 auto args_json = nlohmann::json::parse(
751 func[
"arguments"].get<std::string>(),
nullptr,
false);
752 if (!args_json.is_discarded() && args_json.is_object()) {
753 for (
auto& [key, value] : args_json.items()) {
754 if (value.is_string()) {
755 tool_call.args[
key] = value.get<std::string>();
756 }
else if (value.is_number()) {
757 tool_call.args[
key] = std::to_string(value.get<
double>());
762 agent_response.tool_calls.push_back(tool_call);
767 if (agent_response.text_response.empty() && agent_response.commands.empty() &&
768 agent_response.tool_calls.empty()) {
769 return absl::InternalError(
770 "❌ No valid response extracted from OpenAI\n"
771 " Expected at least one of: text_response, commands, or tool_calls");
774 return agent_response;
OpenAIAIService(const OpenAIConfig &)