diff --git a/packages/uipath-llamaindex/docs/llms_and_embeddings.md b/packages/uipath-llamaindex/docs/llms_and_embeddings.md index be3c147..15e25eb 100644 --- a/packages/uipath-llamaindex/docs/llms_and_embeddings.md +++ b/packages/uipath-llamaindex/docs/llms_and_embeddings.md @@ -49,10 +49,10 @@ print(response) from uipath_llamaindex.llms import UiPathOpenAI, OpenAIModel # Use a specific model -llm = UiPathOpenAI(model=OpenAIModel.GPT_4O_2024_11_20) +llm = UiPathOpenAI(model=OpenAIModel.GPT_4_1_2025_04_14) # Or use a model string directly -llm = UiPathOpenAI(model="gpt-4o-2024-11-20") +llm = UiPathOpenAI(model="gpt-4.1-2025-04-14") ``` ## UiPathOpenAIEmbedding @@ -211,7 +211,7 @@ def add(a: int, b: int) -> int: # Create agent with UiPath LLM agent = ReActAgent( tools=[multiply, add], - llm=UiPathOpenAI(model=OpenAIModel.GPT_4O_2024_11_20)) + llm=UiPathOpenAI(model=OpenAIModel.GPT_4_1_2025_04_14)) async def main(): handler = agent.run("What is 2+(2*4)?") @@ -241,7 +241,7 @@ index = VectorStoreIndex.from_documents( # Create query engine with UiPath LLM query_engine = index.as_query_engine( - llm=UiPathOpenAI(model=OpenAIModel.GPT_4O_2024_11_20) + llm=UiPathOpenAI(model=OpenAIModel.GPT_4_1_2025_04_14) ) response = query_engine.query("What is machine learning?") diff --git a/packages/uipath-llamaindex/pyproject.toml b/packages/uipath-llamaindex/pyproject.toml index 530ecb0..c2e1336 100644 --- a/packages/uipath-llamaindex/pyproject.toml +++ b/packages/uipath-llamaindex/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "uipath-llamaindex" -version = "0.4.0" +version = "0.4.1" description = "UiPath LlamaIndex SDK" readme = { file = "README.md", content-type = "text/markdown" } requires-python = ">=3.11" diff --git a/packages/uipath-llamaindex/samples/action-center-hitl-agent/main.py b/packages/uipath-llamaindex/samples/action-center-hitl-agent/main.py index 78a260c..730a652 100644 --- a/packages/uipath-llamaindex/samples/action-center-hitl-agent/main.py +++ b/packages/uipath-llamaindex/samples/action-center-hitl-agent/main.py @@ -9,7 +9,7 @@ from uipath_llamaindex.models import CreateTaskEvent -llm = OpenAI(model="gpt-4o-mini") +llm = OpenAI() async def may_research_company(ctx: Context, company_name: str) -> bool: diff --git a/packages/uipath-llamaindex/samples/hitl-runtime-demo/main.py b/packages/uipath-llamaindex/samples/hitl-runtime-demo/main.py index 2ef478e..5c5ce50 100644 --- a/packages/uipath-llamaindex/samples/hitl-runtime-demo/main.py +++ b/packages/uipath-llamaindex/samples/hitl-runtime-demo/main.py @@ -11,7 +11,7 @@ from llama_index.llms.openai import OpenAI load_dotenv() -llm = OpenAI(model="gpt-4o-mini") +llm = OpenAI() async def may_research_company(ctx: Context, company_name: str) -> bool: diff --git a/packages/uipath-llamaindex/samples/multi-agent/main.py b/packages/uipath-llamaindex/samples/multi-agent/main.py index 8fff184..1a0ee8c 100644 --- a/packages/uipath-llamaindex/samples/multi-agent/main.py +++ b/packages/uipath-llamaindex/samples/multi-agent/main.py @@ -9,7 +9,7 @@ from uipath_llamaindex.models import InvokeProcessEvent -llm = OpenAI(model="gpt-4o-mini") +llm = OpenAI() async def may_research_company(ctx: Context, company_name: str) -> str: diff --git a/packages/uipath-llamaindex/samples/simple-hitl-agent/main.py b/packages/uipath-llamaindex/samples/simple-hitl-agent/main.py index 79a4b0a..10eed8e 100644 --- a/packages/uipath-llamaindex/samples/simple-hitl-agent/main.py +++ b/packages/uipath-llamaindex/samples/simple-hitl-agent/main.py @@ -6,7 +6,7 @@ ) from llama_index.llms.openai import OpenAI -llm = OpenAI(model="gpt-4o-mini") +llm = OpenAI() async def may_research_company(ctx: Context, company_name: str) -> bool: diff --git a/packages/uipath-llamaindex/samples/simple-remote-mcp-agent/main.py b/packages/uipath-llamaindex/samples/simple-remote-mcp-agent/main.py index 140bf11..96427f3 100644 --- a/packages/uipath-llamaindex/samples/simple-remote-mcp-agent/main.py +++ b/packages/uipath-llamaindex/samples/simple-remote-mcp-agent/main.py @@ -64,7 +64,7 @@ async def process_query(self, ev: UserQueryEvent) -> AgentResponseEvent: self.agent = FunctionAgent( name="UiPath MCP Agent", description="An agent that can interact with MCP tools", - llm=OpenAI(model="gpt-4o"), + llm=OpenAI(), tools=tools, system_prompt="""You are a helpful assistant. You have access to various tools through MCP (Model Context Protocol). diff --git a/packages/uipath-llamaindex/samples/travel-helper-RAG-agent/main.py b/packages/uipath-llamaindex/samples/travel-helper-RAG-agent/main.py index b549795..c3711ea 100644 --- a/packages/uipath-llamaindex/samples/travel-helper-RAG-agent/main.py +++ b/packages/uipath-llamaindex/samples/travel-helper-RAG-agent/main.py @@ -24,7 +24,7 @@ company_policy_files_directory = "sample_data/company_policies" personal_preferences_files_directory = "sample_data/personal_preferences" -llm = UiPathOpenAI(model="gpt-4o-2024-11-20") +llm = UiPathOpenAI() class CustomStartEvent(StartEvent): diff --git a/packages/uipath-llamaindex/src/uipath_llamaindex/llms/_openai.py b/packages/uipath-llamaindex/src/uipath_llamaindex/llms/_openai.py index d118724..e1824c5 100644 --- a/packages/uipath-llamaindex/src/uipath_llamaindex/llms/_openai.py +++ b/packages/uipath-llamaindex/src/uipath_llamaindex/llms/_openai.py @@ -44,7 +44,7 @@ async def handle_async_request(self, request: httpx.Request) -> httpx.Response: class UiPathOpenAI(AzureOpenAI): def __init__( self, - model: str | OpenAIModel = OpenAIModel.GPT_4O_MINI_2024_07_18, + model: str | OpenAIModel = OpenAIModel.GPT_4_1_2025_04_14, api_version: str = "2024-10-21", **kwargs: Any, ): diff --git a/packages/uipath-llamaindex/testcases/quickstart-agent/src/main.py b/packages/uipath-llamaindex/testcases/quickstart-agent/src/main.py index 85a290d..2135a06 100644 --- a/packages/uipath-llamaindex/testcases/quickstart-agent/src/main.py +++ b/packages/uipath-llamaindex/testcases/quickstart-agent/src/main.py @@ -23,7 +23,7 @@ class CritiqueEvent(StopEvent): class JokeFlow(Workflow): - llm = UiPathOpenAI(model="gpt-4o-mini-2024-07-18") + llm = UiPathOpenAI() @step async def generate_joke(self, ev: TopicEvent) -> JokeEvent: diff --git a/packages/uipath-llamaindex/testcases/simple-hitl-agent/src/main.py b/packages/uipath-llamaindex/testcases/simple-hitl-agent/src/main.py index d61619d..e6e6668 100644 --- a/packages/uipath-llamaindex/testcases/simple-hitl-agent/src/main.py +++ b/packages/uipath-llamaindex/testcases/simple-hitl-agent/src/main.py @@ -7,7 +7,7 @@ from uipath_llamaindex.llms import UiPathOpenAI -llm = UiPathOpenAI(model="gpt-4o-mini-2024-07-18") +llm = UiPathOpenAI() async def may_research_company(ctx: Context, company_name: str) -> bool: diff --git a/packages/uipath-llamaindex/uv.lock b/packages/uipath-llamaindex/uv.lock index a67c1fc..acd2f47 100644 --- a/packages/uipath-llamaindex/uv.lock +++ b/packages/uipath-llamaindex/uv.lock @@ -3358,6 +3358,7 @@ dependencies = [ sdist = { url = "https://files.pythonhosted.org/packages/4c/62/de6137c448a4436876e0e57c3cbd08718a7b45c998bdb9e81865ce0c082c/uipath-2.5.tar.gz", hash = "sha256:6ef338d27763e03098d65318fbf126f8a0010f5aa07136ee0486a8283e951ad1", size = 3883309, upload-time = "2026-01-15T07:02:16.8Z" } wheels = [ { url = "https://files.pythonhosted.org/packages/c3/63/8254f1941351b3ee9386d0b073fb3c033e7de9c67ba588a47d5511506e17/uipath-2.5-py3-none-any.whl", hash = "sha256:19f6b5d7408e32e0a9df869dc041a158919758288fdd26bbfe337b9dea12373c", size = 437244, upload-time = "2026-01-15T07:02:14.911Z" }, + { url = "https://files.pythonhosted.org/packages/88/a4/1c81dbee07c01de70281e591c368f832afea956f2e2be7596355716709c8/uipath-2.5.0-py3-none-any.whl", hash = "sha256:7cd24c5048873dd8aaacdb12ec1722be872db9369d410b73ed77d1ce6dfc8e95", size = 437268, upload-time = "2026-01-15T08:39:47.556Z" }, ] [[package]] @@ -3376,7 +3377,7 @@ wheels = [ [[package]] name = "uipath-llamaindex" -version = "0.4.0" +version = "0.4.1" source = { editable = "." } dependencies = [ { name = "aiosqlite" },