chore: add tests

This commit is contained in:
Christopher Gondek 2025-09-01 15:36:32 +02:00
parent 4c3592d7d9
commit cfc83a7f42
4 changed files with 286 additions and 14 deletions

View file

@ -119,7 +119,7 @@ class ConnectorTavily(WebSearchBase, WebCrawlBase, WebScrapeBase):
# Perform actual API call
response = await self.client.search(query=query, max_results=max_results)
logger.info(f"Tavily API response:\n{response}")
logger.info(f"Tavily API search response:\n{response}")
return [
TavilySearchResult(title=result["title"], url=result["url"])
@ -157,6 +157,10 @@ class ConnectorTavily(WebSearchBase, WebCrawlBase, WebScrapeBase):
response = await self.client.extract(
urls=urls, extract_depth="advanced", format="text"
)
# Log the result
logger.info(f"Tavily API extract (crawl) response:\n{response}")
return [
TavilyCrawlResult(url=result["url"], content=result["raw_content"])
for result in response["results"]

View file

@ -4,7 +4,11 @@ import pytest
import logging
from modules.interfaces.interfaceChatModel import ActionResult
from modules.interfaces.interface_web_model import WebSearchRequest
from modules.interfaces.interface_web_model import (
WebSearchRequest,
WebCrawlRequest,
WebScrapeRequest,
)
from modules.connectors.connector_tavily import ConnectorTavily
logger = logging.getLogger(__name__)
@ -13,7 +17,7 @@ logger = logging.getLogger(__name__)
@pytest.mark.asyncio
@pytest.mark.expensive
async def test_tavily_connector_search_test_live_api():
logger.info("Testing Tavliy connector with live API calls")
logger.info("Testing Tavliy connector search with live API calls")
# Test request
request = WebSearchRequest(query="How old is the Earth?", max_results=5)
@ -38,3 +42,67 @@ async def test_tavily_connector_search_test_live_api():
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" - Document Mime Type: {doc.mimeType}")
logger.info(f" - Document Data: {doc.documentData}")
@pytest.mark.asyncio
@pytest.mark.expensive
async def test_tavily_connector_crawl_test_live_api():
logger.info("Testing Tavily connector crawl with live API calls")
# Test request
urls = [
"https://en.wikipedia.org/wiki/Earth",
"https://valueon.ch",
]
request = WebCrawlRequest(urls=urls)
# Tavily instance
connector_tavily = await ConnectorTavily.create()
# Crawl test
action_result = await connector_tavily.crawl_urls(request=request)
# Check results
assert isinstance(action_result, ActionResult)
logger.info("=" * 20)
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info("-" * 10)
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" - Document Mime Type: {doc.mimeType}")
logger.info(f" - Document Data: {doc.documentData}")
@pytest.mark.asyncio
@pytest.mark.expensive
async def test_tavily_connector_scrape_test_live_api():
logger.info("Testing Tavily connector scrape with live API calls")
# Test request with query
request = WebScrapeRequest(query="How old is the Earth?", max_results=3)
# Tavily instance
connector_tavily = await ConnectorTavily.create()
# Scrape test
action_result = await connector_tavily.scrape(request=request)
# Check results
assert isinstance(action_result, ActionResult)
logger.info("=" * 20)
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info("-" * 10)
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" - Document Mime Type: {doc.mimeType}")
logger.info(f" - Document Data: {doc.documentData}")

File diff suppressed because one or more lines are too long

View file

@ -1,11 +1,15 @@
"""Tests for method web.py"""
import json
import logging
import pytest
from unittest.mock import patch
from modules.methods.method_web import MethodWeb
from tests.fixtures.tavily_responses import RESPONSE_HOW_OLD_IS_EARTH_NO_ANSWER
from tests.fixtures.tavily_responses import (
RESPONSE_SEARCH_HOW_OLD_IS_EARTH_NO_ANSWER,
RESPONSE_EXTRACT_HOW_OLD_IS_EARTH_NO_ANSWER,
)
logger = logging.getLogger(__name__)
@ -15,6 +19,9 @@ logger = logging.getLogger(__name__)
async def test_method_web_search_live():
"""Tests method web search with live API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Search Live")
method_web = MethodWeb(serviceCenter=None)
# Actual request
@ -26,29 +33,30 @@ async def test_method_web_search_live():
assert action_result.success
assert len(action_result.documents) > 0
logger.info("=" * 20)
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info("-" * 10)
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" - Document Mime Type: {doc.mimeType}")
logger.info(f" - Document Data: {doc.documentData}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")
@pytest.mark.asyncio
async def test_method_web_search_dummy():
"""Tests method web search with dummy response data - no external API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Search Dummy")
method_web = MethodWeb(serviceCenter=None)
# Mock the Tavily API response
with patch(
"tavily.AsyncTavilyClient.search",
return_value=RESPONSE_HOW_OLD_IS_EARTH_NO_ANSWER,
return_value=RESPONSE_SEARCH_HOW_OLD_IS_EARTH_NO_ANSWER,
) as mock_client:
action_result = await method_web.search(
{"query": "How old is the earth", "maxResults": 5}
@ -59,14 +67,182 @@ async def test_method_web_search_dummy():
assert action_result.success
assert len(action_result.documents) > 0
logger.info("=" * 20)
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info("-" * 10)
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" - Document Mime Type: {doc.mimeType}")
logger.info(f" - Document Data: {doc.documentData}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")
@pytest.mark.asyncio
@pytest.mark.expensive
async def test_method_web_crawl_live():
"""Tests method web crawl with live API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Crawl Live")
method_web = MethodWeb(serviceCenter=None)
# Create mock document data with URLs from search results
search_results_json = {
"documentData": {
"results": [
{"url": "https://en.wikipedia.org/wiki/Age_of_Earth"},
{"url": "https://www.planetary.org/articles/how-old-is-the-earth"},
]
}
}
# Mock the service center methods
with patch.object(method_web, "service") as mock_service:
mock_service.getChatDocumentsFromDocumentList.return_value = [
type("MockDoc", (), {"fileId": "test-file-id"})()
]
mock_service.getFileData.return_value = json.dumps(search_results_json).encode(
"utf-8"
)
# Actual request
action_result = await method_web.crawl({"document": "test-document-ref"})
# Evaluate results
assert action_result.success
assert len(action_result.documents) > 0
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")
@pytest.mark.asyncio
async def test_method_web_crawl_dummy():
"""Tests method web crawl with dummy response data - no external API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Crawl Dummy")
method_web = MethodWeb(serviceCenter=None)
# Create mock document data with URLs from search results
search_results_json = {
"documentData": {
"results": [
{"url": "https://en.wikipedia.org/wiki/Age_of_Earth"},
{"url": "https://www.planetary.org/articles/how-old-is-the-earth"},
]
}
}
# Mock both the service center and Tavily API
with (
patch.object(method_web, "service") as mock_service,
patch(
"tavily.AsyncTavilyClient.extract",
return_value=RESPONSE_EXTRACT_HOW_OLD_IS_EARTH_NO_ANSWER,
) as mock_client,
):
mock_service.getChatDocumentsFromDocumentList.return_value = [
type("MockDoc", (), {"fileId": "test-file-id"})()
]
mock_service.getFileData.return_value = json.dumps(search_results_json).encode(
"utf-8"
)
action_result = await method_web.crawl({"document": "test-document-ref"})
mock_client.assert_called_once()
# Evaluate results
assert action_result.success
assert len(action_result.documents) > 0
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")
@pytest.mark.asyncio
@pytest.mark.expensive
async def test_method_web_scrape_live():
"""Tests method web scrape with live API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Scrape Live")
method_web = MethodWeb(serviceCenter=None)
# Actual request
action_result = await method_web.scrape(
{"query": "How old is the earth", "maxResults": 3}
)
# Evaluate results
assert action_result.success
assert len(action_result.documents) > 0
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")
@pytest.mark.asyncio
async def test_method_web_scrape_dummy():
"""Tests method web scrape with dummy response data - no external API calls."""
logger.info("=" * 50)
logger.info("==> Test: Method Web Scrape Dummy")
method_web = MethodWeb(serviceCenter=None)
# Mock both Tavily API responses (search + extract)
with (
patch(
"tavily.AsyncTavilyClient.search",
return_value=RESPONSE_SEARCH_HOW_OLD_IS_EARTH_NO_ANSWER,
) as mock_search,
patch(
"tavily.AsyncTavilyClient.extract",
return_value=RESPONSE_EXTRACT_HOW_OLD_IS_EARTH_NO_ANSWER,
) as mock_extract,
):
action_result = await method_web.scrape(
{"query": "How old is the earth", "maxResults": 3}
)
mock_search.assert_called_once()
mock_extract.assert_called_once()
# Evaluate results
assert action_result.success
assert len(action_result.documents) > 0
logger.info(f"Action result success status: {action_result.success}")
logger.info(f"Action result error: {action_result.error}")
logger.info(f"Action result label: {action_result.resultLabel}")
logger.info("Documents:")
for doc in action_result.documents:
logger.info(f" - Document Name: {doc.documentName}")
logger.info(f" --> Document Mime Type: {doc.mimeType}")
logger.info(f" --> Document Data: {doc.documentData}")