928 lines
42 KiB
Python
928 lines
42 KiB
Python
# Copyright (c) 2025 Patrick Motsch
|
|
# All rights reserved.
|
|
"""Connector for SharePoint operations using Microsoft Graph API."""
|
|
|
|
import logging
|
|
import aiohttp
|
|
import asyncio
|
|
import time
|
|
from typing import Dict, Any, List, Optional
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
# Cache for discoverSites() to avoid hitting Graph API on every folder-options call (e.g. when UI loads site list).
|
|
# Key: token prefix (per user), Value: (expiry_ts, sites). TTL 5 minutes.
|
|
_discoverSitesCache: Dict[str, tuple] = {}
|
|
_DISCOVER_SITES_TTL_SEC = 300
|
|
|
|
|
|
class SharepointService:
|
|
"""SharePoint connector using Microsoft Graph API for reliable authentication."""
|
|
|
|
def __init__(self, serviceCenter=None):
|
|
"""Initialize SharePoint service without access token.
|
|
|
|
Args:
|
|
serviceCenter: Service center instance for accessing other services
|
|
|
|
Use setAccessTokenFromConnection() method to configure the access token before making API calls.
|
|
"""
|
|
self.services = serviceCenter
|
|
self.accessToken = None
|
|
self.baseUrl = "https://graph.microsoft.com/v1.0"
|
|
|
|
def setAccessTokenFromConnection(self, userConnection) -> bool:
|
|
"""Set access token from UserConnection.
|
|
|
|
Args:
|
|
userConnection: UserConnection object or dict containing token information
|
|
|
|
Returns:
|
|
bool: True if token was set successfully, False otherwise
|
|
"""
|
|
try:
|
|
if not userConnection:
|
|
logger.error("UserConnection is required to set access token")
|
|
return False
|
|
|
|
# Handle both dict and UserConnection object
|
|
if isinstance(userConnection, dict):
|
|
connectionId = userConnection.get('id')
|
|
else:
|
|
connectionId = getattr(userConnection, 'id', None)
|
|
|
|
if not connectionId:
|
|
logger.error("UserConnection must have an 'id' field")
|
|
return False
|
|
|
|
# Get a fresh token for this specific connection via security service
|
|
if not self.services:
|
|
logger.error("Service center not available for token access")
|
|
return False
|
|
|
|
token = self.services.security.getFreshToken(connectionId)
|
|
if not token:
|
|
logger.error(f"No token found for connection {connectionId}")
|
|
return False
|
|
|
|
self.accessToken = token.tokenAccess
|
|
logger.info(f"Access token set for connection {connectionId}")
|
|
return True
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error setting access token: {str(e)}")
|
|
return False
|
|
|
|
async def _makeGraphApiCall(self, endpoint: str, method: str = "GET", data: bytes = None) -> Dict[str, Any]:
|
|
"""Make a Microsoft Graph API call with proper error handling."""
|
|
try:
|
|
if self.accessToken is None:
|
|
logger.error("Access token is not set. Please call setAccessTokenFromConnection() before using the SharePoint service.")
|
|
return {"error": "Access token is not set. Please call setAccessTokenFromConnection() before using the SharePoint service."}
|
|
|
|
headers = {
|
|
"Authorization": f"Bearer {self.accessToken}",
|
|
"Content-Type": "application/json" if data and method != "PUT" else "application/octet-stream" if data else "application/json"
|
|
}
|
|
|
|
# Remove leading slash from endpoint to avoid double slash
|
|
cleanEndpoint = endpoint.lstrip('/')
|
|
url = f"{self.baseUrl}/{cleanEndpoint}"
|
|
logger.debug(f"Making Graph API call: {method} {url}")
|
|
|
|
timeout = aiohttp.ClientTimeout(total=30)
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
if method == "GET":
|
|
async with session.get(url, headers=headers) as response:
|
|
if response.status == 200:
|
|
return await response.json()
|
|
else:
|
|
error_text = await response.text()
|
|
logger.error(f"Graph API call failed: {response.status} - {error_text}")
|
|
return {"error": f"API call failed: {response.status} - {error_text}"}
|
|
|
|
elif method == "PUT":
|
|
async with session.put(url, headers=headers, data=data) as response:
|
|
if response.status in [200, 201]:
|
|
return await response.json()
|
|
else:
|
|
error_text = await response.text()
|
|
logger.error(f"Graph API call failed: {response.status} - {error_text}")
|
|
return {"error": f"API call failed: {response.status} - {error_text}"}
|
|
|
|
elif method == "POST":
|
|
async with session.post(url, headers=headers, data=data) as response:
|
|
if response.status in [200, 201]:
|
|
return await response.json()
|
|
else:
|
|
error_text = await response.text()
|
|
logger.error(f"Graph API call failed: {response.status} - {error_text}")
|
|
return {"error": f"API call failed: {response.status} - {error_text}"}
|
|
|
|
except asyncio.TimeoutError:
|
|
logger.error(f"Graph API call timed out after 30 seconds: {endpoint}")
|
|
return {"error": f"API call timed out after 30 seconds: {endpoint}"}
|
|
except Exception as e:
|
|
logger.error(f"Error making Graph API call: {str(e)}")
|
|
return {"error": f"Error making Graph API call: {str(e)}"}
|
|
|
|
async def discoverSites(self) -> List[Dict[str, Any]]:
|
|
"""Discover all SharePoint sites accessible to the user."""
|
|
try:
|
|
result = await self._makeGraphApiCall("sites?search=*")
|
|
|
|
if "error" in result:
|
|
logger.error(f"Error discovering SharePoint sites: {result['error']}")
|
|
return []
|
|
|
|
sites = result.get("value", [])
|
|
logger.info(f"Discovered {len(sites)} SharePoint sites")
|
|
|
|
processedSites = []
|
|
for site in sites:
|
|
siteInfo = {
|
|
"id": site.get("id"),
|
|
"displayName": site.get("displayName"),
|
|
"name": site.get("name"),
|
|
"webUrl": site.get("webUrl"),
|
|
"description": site.get("description"),
|
|
"createdDateTime": site.get("createdDateTime"),
|
|
"lastModifiedDateTime": site.get("lastModifiedDateTime")
|
|
}
|
|
processedSites.append(siteInfo)
|
|
logger.debug(f"Site: {siteInfo['displayName']} - {siteInfo['webUrl']}")
|
|
|
|
return processedSites
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error discovering SharePoint sites: {str(e)}")
|
|
return []
|
|
|
|
async def findSiteByName(self, siteName: str) -> Optional[Dict[str, Any]]:
|
|
"""Find a specific SharePoint site by name using direct Graph API call."""
|
|
try:
|
|
# Try to get the site directly by name using Graph API
|
|
endpoint = f"sites/{siteName}"
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if result and "error" not in result:
|
|
siteInfo = {
|
|
"id": result.get("id"),
|
|
"displayName": result.get("displayName"),
|
|
"name": result.get("name"),
|
|
"webUrl": result.get("webUrl"),
|
|
"description": result.get("description"),
|
|
"createdDateTime": result.get("createdDateTime"),
|
|
"lastModifiedDateTime": result.get("lastModifiedDateTime")
|
|
}
|
|
logger.info(f"Found site directly: {siteInfo['displayName']} - {siteInfo['webUrl']}")
|
|
return siteInfo
|
|
|
|
except Exception as e:
|
|
logger.debug(f"Direct site lookup failed for '{siteName}': {str(e)}")
|
|
|
|
# Fallback to discovery if direct lookup fails
|
|
logger.info(f"Direct lookup failed, trying discovery for site: {siteName}")
|
|
sites = await self.discoverSites()
|
|
if not sites:
|
|
logger.warning("No sites discovered")
|
|
return None
|
|
|
|
logger.info(f"Discovered {len(sites)} SharePoint sites:")
|
|
for site in sites:
|
|
logger.info(f" - {site.get('displayName', 'Unknown')} (ID: {site.get('id', 'Unknown')})")
|
|
|
|
# Try exact match first
|
|
for site in sites:
|
|
if site.get("displayName", "").strip().lower() == siteName.strip().lower():
|
|
logger.info(f"Found exact match: {site.get('displayName')}")
|
|
return site
|
|
|
|
# Try partial match
|
|
for site in sites:
|
|
if siteName.lower() in site.get("displayName", "").lower():
|
|
logger.info(f"Found partial match: {site.get('displayName')}")
|
|
return site
|
|
|
|
logger.warning(f"No site found matching: {siteName}")
|
|
return None
|
|
|
|
async def findSiteByWebUrl(self, webUrl: str) -> Optional[Dict[str, Any]]:
|
|
"""Find a SharePoint site using its web URL (useful for guest sites)."""
|
|
try:
|
|
# Use the web URL format: sites/{hostname}:/sites/{site-path}
|
|
# Extract hostname and site path from the web URL
|
|
if not webUrl.startswith("https://"):
|
|
webUrl = f"https://{webUrl}"
|
|
|
|
# Parse the URL to extract hostname and site path
|
|
from urllib.parse import urlparse
|
|
parsed = urlparse(webUrl)
|
|
hostname = parsed.hostname
|
|
pathParts = parsed.path.strip('/').split('/')
|
|
|
|
if len(pathParts) >= 2 and pathParts[0] == 'sites':
|
|
sitePath = '/'.join(pathParts[1:]) # Everything after 'sites/'
|
|
else:
|
|
logger.error(f"Invalid SharePoint URL format: {webUrl}")
|
|
return None
|
|
|
|
endpoint = f"sites/{hostname}:/sites/{sitePath}"
|
|
logger.debug(f"Trying web URL format: {endpoint}")
|
|
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if result and "error" not in result:
|
|
siteInfo = {
|
|
"id": result.get("id"),
|
|
"displayName": result.get("displayName"),
|
|
"name": result.get("name"),
|
|
"webUrl": result.get("webUrl"),
|
|
"description": result.get("description"),
|
|
"createdDateTime": result.get("createdDateTime"),
|
|
"lastModifiedDateTime": result.get("lastModifiedDateTime")
|
|
}
|
|
logger.info(f"Found site by web URL: {siteInfo['displayName']} - {siteInfo['webUrl']} (ID: {siteInfo['id']})")
|
|
return siteInfo
|
|
else:
|
|
logger.warning(f"Site not found using web URL: {webUrl}")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error finding site by web URL: {str(e)}")
|
|
return None
|
|
|
|
async def findSiteByUrl(self, hostname: str, sitePath: str) -> Optional[Dict[str, Any]]:
|
|
"""Find a SharePoint site using the site URL format."""
|
|
try:
|
|
# For guest sites, try different URL formats
|
|
urlFormats = [
|
|
f"sites/{hostname}:/sites/{sitePath}", # Standard format
|
|
f"sites/{hostname}:/sites/{sitePath}/", # With trailing slash
|
|
f"sites/{hostname}:/sites/{sitePath.lower()}", # Lowercase
|
|
f"sites/{hostname}:/sites/{sitePath.lower()}/", # Lowercase with slash
|
|
]
|
|
|
|
for endpoint in urlFormats:
|
|
logger.debug(f"Trying URL format: {endpoint}")
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if result and "error" not in result:
|
|
siteInfo = {
|
|
"id": result.get("id"),
|
|
"displayName": result.get("displayName"),
|
|
"name": result.get("name"),
|
|
"webUrl": result.get("webUrl"),
|
|
"description": result.get("description"),
|
|
"createdDateTime": result.get("createdDateTime"),
|
|
"lastModifiedDateTime": result.get("lastModifiedDateTime")
|
|
}
|
|
logger.info(f"Found site by URL: {siteInfo['displayName']} - {siteInfo['webUrl']} (ID: {siteInfo['id']})")
|
|
return siteInfo
|
|
else:
|
|
logger.debug(f"URL format failed: {endpoint} - {result.get('error', 'Unknown error')}")
|
|
|
|
logger.warning(f"Site not found using any URL format for: {hostname}:/sites/{sitePath}")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error finding site by URL: {str(e)}")
|
|
return None
|
|
|
|
async def getFolderByPath(self, siteId: str, folderPath: str) -> Optional[Dict[str, Any]]:
|
|
"""Get folder information by path within a site."""
|
|
try:
|
|
# Clean the path
|
|
cleanPath = folderPath.lstrip('/')
|
|
|
|
# If path is empty, get root directly
|
|
if not cleanPath:
|
|
endpoint = f"sites/{siteId}/drive/root"
|
|
else:
|
|
endpoint = f"sites/{siteId}/drive/root:/{cleanPath}"
|
|
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.warning(f"Folder not found at path {folderPath}: {result['error']}")
|
|
return None
|
|
|
|
return result
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error getting folder by path: {str(e)}")
|
|
return None
|
|
|
|
async def uploadFile(self, siteId: str, folderPath: str, fileName: str, content: bytes) -> Dict[str, Any]:
|
|
"""Upload a file to SharePoint."""
|
|
try:
|
|
# Clean the path
|
|
cleanPath = folderPath.lstrip('/')
|
|
uploadPath = f"{cleanPath.rstrip('/')}/{fileName}"
|
|
endpoint = f"sites/{siteId}/drive/root:/{uploadPath}:/content"
|
|
|
|
logger.info(f"Uploading file to: {endpoint}")
|
|
|
|
result = await self._makeGraphApiCall(endpoint, method="PUT", data=content)
|
|
|
|
if "error" in result:
|
|
logger.error(f"Upload failed: {result['error']}")
|
|
return result
|
|
|
|
logger.info(f"File uploaded successfully: {fileName}")
|
|
return result
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error uploading file: {str(e)}")
|
|
return {"error": f"Error uploading file: {str(e)}"}
|
|
|
|
async def downloadFile(self, siteId: str, fileId: str) -> Optional[bytes]:
|
|
"""Download a file from SharePoint."""
|
|
try:
|
|
if self.accessToken is None:
|
|
logger.error("Access token is not set. Please call setAccessTokenFromConnection() before using the SharePoint service.")
|
|
return None
|
|
|
|
endpoint = f"sites/{siteId}/drive/items/{fileId}/content"
|
|
|
|
headers = {"Authorization": f"Bearer {self.accessToken}"}
|
|
timeout = aiohttp.ClientTimeout(total=30)
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
async with session.get(f"{self.baseUrl}/{endpoint}", headers=headers) as response:
|
|
if response.status == 200:
|
|
return await response.read()
|
|
else:
|
|
logger.error(f"Download failed: {response.status}")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error downloading file: {str(e)}")
|
|
return None
|
|
|
|
async def listFolderContents(self, siteId: str, folderPath: str = "") -> List[Dict[str, Any]]:
|
|
"""List contents of a folder."""
|
|
try:
|
|
if not folderPath or folderPath == "/":
|
|
endpoint = f"sites/{siteId}/drive/root/children"
|
|
else:
|
|
cleanPath = folderPath.lstrip('/')
|
|
endpoint = f"sites/{siteId}/drive/root:/{cleanPath}:/children"
|
|
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.warning(f"Failed to list folder contents: {result['error']}")
|
|
return None
|
|
|
|
items = result.get("value", [])
|
|
processedItems = []
|
|
|
|
for item in items:
|
|
# Determine if it's a folder or file
|
|
isFolder = 'folder' in item
|
|
|
|
itemInfo = {
|
|
"id": item.get("id"),
|
|
"name": item.get("name"),
|
|
"type": "folder" if isFolder else "file",
|
|
"size": item.get("size", 0),
|
|
"createdDateTime": item.get("createdDateTime"),
|
|
"lastModifiedDateTime": item.get("lastModifiedDateTime"),
|
|
"webUrl": item.get("webUrl")
|
|
}
|
|
|
|
if "file" in item:
|
|
itemInfo["mimeType"] = item["file"].get("mimeType")
|
|
itemInfo["downloadUrl"] = item.get("@microsoft.graph.downloadUrl")
|
|
|
|
if "folder" in item:
|
|
itemInfo["childCount"] = item["folder"].get("childCount", 0)
|
|
|
|
processedItems.append(itemInfo)
|
|
|
|
return processedItems
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error listing folder contents: {str(e)}")
|
|
return []
|
|
|
|
async def searchFiles(self, siteId: str, query: str) -> List[Dict[str, Any]]:
|
|
"""Search for files in a site."""
|
|
try:
|
|
searchQuery = query.replace("'", "''") # Escape single quotes for OData
|
|
endpoint = f"sites/{siteId}/drive/root/search(q='{searchQuery}')"
|
|
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.warning(f"Search failed: {result['error']}")
|
|
return []
|
|
|
|
items = result.get("value", [])
|
|
processedItems = []
|
|
|
|
for item in items:
|
|
isFolder = 'folder' in item
|
|
|
|
itemInfo = {
|
|
"id": item.get("id"),
|
|
"name": item.get("name"),
|
|
"type": "folder" if isFolder else "file",
|
|
"size": item.get("size", 0),
|
|
"createdDateTime": item.get("createdDateTime"),
|
|
"lastModifiedDateTime": item.get("lastModifiedDateTime"),
|
|
"webUrl": item.get("webUrl"),
|
|
"parentPath": item.get("parentReference", {}).get("path", "")
|
|
}
|
|
|
|
if "file" in item:
|
|
itemInfo["mimeType"] = item["file"].get("mimeType")
|
|
itemInfo["downloadUrl"] = item.get("@microsoft.graph.downloadUrl")
|
|
|
|
processedItems.append(itemInfo)
|
|
|
|
return processedItems
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error searching files: {str(e)}")
|
|
return []
|
|
|
|
async def copyFileAsync(self, siteId: str, sourceFolder: str, sourceFile: str, destFolder: str, destFile: str) -> None:
|
|
"""Copy a file from source to destination folder (like original synchronizer)."""
|
|
try:
|
|
# First, download the source file
|
|
sourcePath = f"{sourceFolder}/{sourceFile}"
|
|
fileContent = await self.downloadFileByPath(siteId=siteId, filePath=sourcePath)
|
|
|
|
if not fileContent:
|
|
raise Exception(f"Failed to download source file: {sourcePath}")
|
|
|
|
# Upload to destination
|
|
await self.uploadFile(
|
|
siteId=siteId,
|
|
folderPath=destFolder,
|
|
fileName=destFile,
|
|
content=fileContent
|
|
)
|
|
|
|
logger.info(f"File copied: {sourceFile} -> {destFile}")
|
|
|
|
except Exception as e:
|
|
# Provide more specific error information
|
|
errorMsg = str(e)
|
|
if "itemNotFound" in errorMsg or "404" in errorMsg:
|
|
raise Exception(f"Source file not found (404): {sourcePath} - {errorMsg}")
|
|
else:
|
|
raise Exception(f"Error copying file: {errorMsg}")
|
|
|
|
async def downloadFileByPath(self, siteId: str, filePath: str) -> Optional[bytes]:
|
|
"""Download a file by its path within a site."""
|
|
try:
|
|
if self.accessToken is None:
|
|
logger.error("Access token is not set. Please call setAccessTokenFromConnection() before using the SharePoint service.")
|
|
return None
|
|
|
|
# Clean the path
|
|
cleanPath = filePath.strip('/')
|
|
endpoint = f"sites/{siteId}/drive/root:/{cleanPath}:/content"
|
|
|
|
# Use direct HTTP call for file downloads (binary content)
|
|
headers = {
|
|
"Authorization": f"Bearer {self.accessToken}",
|
|
}
|
|
|
|
# Remove leading slash from endpoint to avoid double slash
|
|
cleanEndpoint = endpoint.lstrip('/')
|
|
url = f"{self.baseUrl}/{cleanEndpoint}"
|
|
logger.debug(f"Downloading file: GET {url}")
|
|
|
|
timeout = aiohttp.ClientTimeout(total=30)
|
|
|
|
async with aiohttp.ClientSession(timeout=timeout) as session:
|
|
async with session.get(url, headers=headers) as response:
|
|
if response.status == 200:
|
|
return await response.read()
|
|
else:
|
|
error_text = await response.text()
|
|
logger.error(f"File download failed: {response.status} - {error_text}")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error downloading file by path: {str(e)}")
|
|
return None
|
|
|
|
async def _getItemById(self, siteId: str, driveId: str, itemId: str) -> Optional[Dict[str, Any]]:
|
|
"""Verify that an item exists by getting it by ID.
|
|
|
|
Args:
|
|
siteId: SharePoint site ID
|
|
driveId: Drive ID (document library)
|
|
itemId: Item ID to verify
|
|
|
|
Returns:
|
|
Item dictionary if found, None otherwise
|
|
"""
|
|
try:
|
|
endpoint = f"sites/{siteId}/drives/{driveId}/items/{itemId}"
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.warning(f"Item {itemId} not found: {result['error']}")
|
|
return None
|
|
|
|
return result
|
|
|
|
except Exception as e:
|
|
logger.warning(f"Error verifying item {itemId}: {str(e)}")
|
|
return None
|
|
|
|
async def _findDriveForItem(self, siteId: str, itemId: str) -> Optional[str]:
|
|
"""Find which drive contains a specific item by trying to get it from all drives.
|
|
|
|
Args:
|
|
siteId: SharePoint site ID
|
|
itemId: Item ID to find
|
|
|
|
Returns:
|
|
Drive ID if found, None otherwise
|
|
"""
|
|
try:
|
|
# Get all drives for the site
|
|
endpoint = f"sites/{siteId}/drives"
|
|
drivesResult = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in drivesResult:
|
|
logger.warning(f"Could not get drives for site {siteId}: {drivesResult['error']}")
|
|
return None
|
|
|
|
drives = drivesResult.get("value", [])
|
|
if not drives:
|
|
logger.warning(f"No drives found for site {siteId}")
|
|
return None
|
|
|
|
# Try to find the item in each drive
|
|
for drive in drives:
|
|
driveId = drive.get("id")
|
|
if not driveId:
|
|
continue
|
|
|
|
itemInfo = await self._getItemById(siteId, driveId, itemId)
|
|
if itemInfo:
|
|
logger.info(f"Found item {itemId} in drive {drive.get('name', driveId)}")
|
|
return driveId
|
|
|
|
logger.warning(f"Item {itemId} not found in any drive for site {siteId}")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.warning(f"Error finding drive for item {itemId}: {str(e)}")
|
|
return None
|
|
|
|
async def getFolderUsageAnalytics(self, siteId: str, driveId: str, itemId: str, startDateTime: Optional[str] = None, endDateTime: Optional[str] = None, interval: str = "day") -> Dict[str, Any]:
|
|
"""Get usage analytics for a folder or file.
|
|
|
|
Args:
|
|
siteId: SharePoint site ID
|
|
driveId: Drive ID (document library)
|
|
itemId: Folder or file item ID
|
|
startDateTime: Start date/time in ISO format (e.g., "2025-11-01T00:00:00Z"). If None, uses 30 days ago.
|
|
endDateTime: End date/time in ISO format (e.g., "2025-11-30T23:59:59Z"). If None, uses current time.
|
|
interval: Time interval for grouping activities. Options: "day", "week", "month". Default: "day"
|
|
|
|
Returns:
|
|
Dictionary containing analytics data with activities grouped by interval.
|
|
If analytics are not available (404), returns empty analytics structure instead of error.
|
|
"""
|
|
try:
|
|
from datetime import datetime, timedelta, timezone
|
|
|
|
# Set default time range if not provided (last 30 days)
|
|
if not endDateTime:
|
|
endDateTime = datetime.now(timezone.utc).isoformat().replace('+00:00', 'Z')
|
|
if not startDateTime:
|
|
startDate = datetime.now(timezone.utc) - timedelta(days=30)
|
|
startDateTime = startDate.isoformat().replace('+00:00', 'Z')
|
|
|
|
# Build endpoint with query parameters
|
|
endpoint = f"sites/{siteId}/drives/{driveId}/items/{itemId}/getActivitiesByInterval"
|
|
endpoint += f"?startDateTime={startDateTime}&endDateTime={endDateTime}&interval={interval}"
|
|
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
errorMsg = result.get('error', '')
|
|
# Check if it's a 404 error
|
|
if isinstance(errorMsg, str) and '404' in errorMsg:
|
|
# Verify if the item exists - first try with current driveId
|
|
itemInfo = await self._getItemById(siteId, driveId, itemId)
|
|
|
|
# If not found, try to find the correct drive for this item
|
|
if not itemInfo:
|
|
logger.info(f"Item {itemId} not found in drive {driveId}, searching for correct drive")
|
|
correctDriveId = await self._findDriveForItem(siteId, itemId)
|
|
if correctDriveId and correctDriveId != driveId:
|
|
logger.info(f"Found item in different drive {correctDriveId}, retrying analytics call")
|
|
# Retry with correct drive
|
|
endpoint = f"sites/{siteId}/drives/{correctDriveId}/items/{itemId}/getActivitiesByInterval"
|
|
endpoint += f"?startDateTime={startDateTime}&endDateTime={endDateTime}&interval={interval}"
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" not in result:
|
|
logger.info(f"Successfully retrieved analytics using correct drive {correctDriveId}")
|
|
return result
|
|
# If still error, continue with original error handling
|
|
itemInfo = await self._getItemById(siteId, correctDriveId, itemId)
|
|
|
|
if itemInfo:
|
|
# Item exists but analytics are not available - return empty analytics
|
|
logger.warning(f"Usage analytics not available for item {itemId} (item exists but has no activity data or analytics not supported)")
|
|
return {
|
|
"value": [],
|
|
"note": "No analytics data available for this item. The item exists but may not have activity data or analytics may not be supported for this item type."
|
|
}
|
|
else:
|
|
# Item doesn't exist
|
|
logger.error(f"Item {itemId} not found when trying to get usage analytics")
|
|
return result
|
|
else:
|
|
# Other error
|
|
logger.error(f"Error getting usage analytics: {result['error']}")
|
|
return result
|
|
|
|
logger.info(f"Retrieved usage analytics for item {itemId} with interval {interval}")
|
|
return result
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error getting folder usage analytics: {str(e)}")
|
|
return {"error": f"Error getting folder usage analytics: {str(e)}"}
|
|
|
|
async def getDriveId(self, siteId: str, driveName: Optional[str] = None) -> Optional[str]:
|
|
"""Get drive ID for a site. If driveName is provided, finds the specific drive, otherwise returns the default drive.
|
|
|
|
Args:
|
|
siteId: SharePoint site ID
|
|
driveName: Optional drive name (document library name). If None, returns default drive.
|
|
|
|
Returns:
|
|
Drive ID string or None if not found
|
|
"""
|
|
try:
|
|
endpoint = f"sites/{siteId}/drives"
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.error(f"Error getting drives: {result['error']}")
|
|
return None
|
|
|
|
drives = result.get("value", [])
|
|
|
|
if not driveName:
|
|
# Return default drive (usually the first one or the one named "Documents")
|
|
for drive in drives:
|
|
if drive.get("name") == "Documents" or drive.get("name") == "Shared Documents":
|
|
logger.info(f"Found default drive: {drive.get('name')} (ID: {drive.get('id')})")
|
|
return drive.get("id")
|
|
# If no Documents drive found, return first drive
|
|
if drives:
|
|
logger.info(f"Using first drive: {drives[0].get('name')} (ID: {drives[0].get('id')})")
|
|
return drives[0].get("id")
|
|
return None
|
|
|
|
# Find specific drive by name
|
|
for drive in drives:
|
|
if drive.get("name", "").lower() == driveName.lower():
|
|
logger.info(f"Found drive '{driveName}': {drive.get('id')}")
|
|
return drive.get("id")
|
|
|
|
logger.warning(f"Drive '{driveName}' not found")
|
|
return None
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error getting drive ID: {str(e)}")
|
|
return None
|
|
|
|
def extractSiteFromStandardPath(self, pathQuery: str) -> Optional[Dict[str, str]]:
|
|
"""
|
|
Extract site name from Microsoft-standard server-relative path:
|
|
/sites/company-share/Freigegebene Dokumente/...
|
|
|
|
Returns dict with keys: siteName, innerPath (no leading slash) on success, else None.
|
|
"""
|
|
try:
|
|
if not pathQuery or not pathQuery.startswith('/sites/'):
|
|
return None
|
|
|
|
# Remove leading /sites/ prefix
|
|
remainder = pathQuery[7:] # len('/sites/') = 7
|
|
|
|
# Split on first '/' to get site name
|
|
if '/' not in remainder:
|
|
# Only site name, no inner path
|
|
return {"siteName": remainder, "innerPath": ""}
|
|
|
|
siteName, inner = remainder.split('/', 1)
|
|
siteName = siteName.strip()
|
|
innerPath = inner.strip()
|
|
|
|
if not siteName:
|
|
return None
|
|
|
|
return {"siteName": siteName, "innerPath": innerPath}
|
|
except Exception as e:
|
|
logger.error(f"Error extracting site from standard path '{pathQuery}': {str(e)}")
|
|
return None
|
|
|
|
async def getSiteByStandardPath(self, sitePath: str, allSites: Optional[List[Dict[str, Any]]] = None) -> Optional[Dict[str, Any]]:
|
|
"""
|
|
Get SharePoint site directly by Microsoft-standard path (/sites/SiteName)
|
|
without loading all sites. Uses hostname from root site (single API call).
|
|
|
|
Parameters:
|
|
sitePath (str): Site path like 'company-share' (without /sites/ prefix)
|
|
allSites (Optional[List[Dict]]): Pre-discovered sites list (optional, for optimization)
|
|
|
|
Returns:
|
|
Optional[Dict[str, Any]]: Site information if found, None otherwise
|
|
"""
|
|
try:
|
|
from urllib.parse import urlparse
|
|
hostname = None
|
|
|
|
# Get hostname - priority order: allSites > root site API > full discovery (fallback)
|
|
if allSites and len(allSites) > 0:
|
|
webUrl = allSites[0].get("webUrl", "")
|
|
hostname = urlparse(webUrl).hostname if webUrl else None
|
|
|
|
if not hostname:
|
|
# Get hostname from root site (single efficient API call instead of discovering all sites)
|
|
rootSite = await self._makeGraphApiCall("sites/root")
|
|
if rootSite and "webUrl" in rootSite and "error" not in rootSite:
|
|
hostname = urlparse(rootSite.get("webUrl", "")).hostname
|
|
logger.debug(f"Got hostname '{hostname}' from root site (efficient)")
|
|
|
|
if not hostname:
|
|
# Fallback: discover all sites (expensive, avoid if possible)
|
|
logger.warning("Could not get hostname from root site, falling back to full site discovery")
|
|
minimalSites = await self.discoverSites()
|
|
if not minimalSites:
|
|
logger.warning("No sites available to extract hostname")
|
|
return None
|
|
hostname = urlparse(minimalSites[0].get("webUrl", "")).hostname
|
|
|
|
if not hostname:
|
|
logger.warning("Could not extract hostname from site")
|
|
return None
|
|
|
|
logger.info(f"Extracted hostname '{hostname}' from first site, now getting site by path: {sitePath}")
|
|
|
|
# Get site directly using hostname + path
|
|
endpoint = f"sites/{hostname}:/sites/{sitePath}"
|
|
result = await self._makeGraphApiCall(endpoint)
|
|
|
|
if "error" in result:
|
|
logger.warning(f"Could not get site directly by path '{sitePath}': {result['error']}")
|
|
return None
|
|
|
|
siteInfo = {
|
|
"id": result.get("id"),
|
|
"displayName": result.get("displayName"),
|
|
"name": result.get("name"),
|
|
"webUrl": result.get("webUrl"),
|
|
"description": result.get("description"),
|
|
"createdDateTime": result.get("createdDateTime"),
|
|
"lastModifiedDateTime": result.get("lastModifiedDateTime")
|
|
}
|
|
|
|
logger.info(f"Successfully got site by standard path: {siteInfo['displayName']} (ID: {siteInfo['id']})")
|
|
return siteInfo
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error getting site by standard path '{sitePath}': {str(e)}")
|
|
return None
|
|
|
|
def filterSitesByHint(self, sites: List[Dict[str, Any]], siteHint: str) -> List[Dict[str, Any]]:
|
|
"""Filter discovered sites by a human-entered site hint (case-insensitive substring)."""
|
|
try:
|
|
if not siteHint:
|
|
return sites
|
|
hint = siteHint.strip().lower()
|
|
filtered: List[Dict[str, Any]] = []
|
|
for site in sites:
|
|
name = (site.get("displayName") or "").lower()
|
|
webUrl = (site.get("webUrl") or "").lower()
|
|
if hint in name or hint in webUrl:
|
|
filtered.append(site)
|
|
return filtered if filtered else sites
|
|
except Exception as e:
|
|
logger.error(f"Error filtering sites by hint '{siteHint}': {str(e)}")
|
|
return sites
|
|
|
|
async def resolveSitesFromPathQuery(self, pathQuery: str, allSites: Optional[List[Dict[str, Any]]] = None) -> List[Dict[str, Any]]:
|
|
"""
|
|
Resolve sites from pathQuery. Handles both Microsoft-standard paths (/sites/SiteName/...)
|
|
and regular paths. Returns list of matching sites.
|
|
|
|
Parameters:
|
|
pathQuery (str): Path query string (e.g., /sites/SiteName/FolderPath)
|
|
allSites (Optional[List[Dict]]): Pre-discovered sites list (optional, for optimization)
|
|
|
|
Returns:
|
|
List[Dict[str, Any]]: List of matching sites
|
|
"""
|
|
try:
|
|
# If pathQuery starts with Microsoft-standard /sites/, try to get site directly
|
|
if pathQuery.startswith('/sites/'):
|
|
parsedPath = self.extractSiteFromStandardPath(pathQuery)
|
|
if parsedPath:
|
|
siteName = parsedPath.get("siteName")
|
|
directSite = await self.getSiteByStandardPath(siteName, allSites)
|
|
if directSite:
|
|
logger.info(f"Got site directly by standard path - no need to discover all sites")
|
|
return [directSite]
|
|
else:
|
|
logger.warning(f"Could not get site directly, falling back to site discovery")
|
|
|
|
# If we didn't get the site directly, use discovery and filtering
|
|
if not allSites:
|
|
allSites = await self.discoverSites()
|
|
if not allSites:
|
|
logger.warning("No SharePoint sites found or accessible")
|
|
return []
|
|
|
|
# If pathQuery starts with Microsoft-standard /sites/, extract site name and filter
|
|
if pathQuery.startswith('/sites/'):
|
|
parsedPath = self.extractSiteFromStandardPath(pathQuery)
|
|
if parsedPath:
|
|
siteName = parsedPath.get("siteName")
|
|
sites = self.filterSitesByHint(allSites, siteName)
|
|
if not sites:
|
|
logger.warning(f"No SharePoint site found matching '{siteName}'")
|
|
return []
|
|
logger.info(f"Filtered to site(s) matching '{siteName}': {[s['displayName'] for s in sites]}")
|
|
return sites
|
|
else:
|
|
return allSites
|
|
else:
|
|
return allSites
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error resolving sites from pathQuery '{pathQuery}': {str(e)}")
|
|
return []
|
|
|
|
def validatePathQuery(self, pathQuery: str) -> tuple[bool, Optional[str]]:
|
|
"""
|
|
Validate pathQuery format. Returns (isValid, errorMessage).
|
|
|
|
Parameters:
|
|
pathQuery (str): Path query to validate
|
|
|
|
Returns:
|
|
tuple[bool, Optional[str]]: (True, None) if valid, (False, errorMessage) if invalid
|
|
"""
|
|
try:
|
|
if not pathQuery or pathQuery.strip() == "" or pathQuery.strip() == "*":
|
|
return False, "pathQuery cannot be empty or '*'"
|
|
|
|
if not pathQuery.startswith('/'):
|
|
return False, "pathQuery must start with '/' and include site name with Microsoft-standard syntax /sites/<SiteName>/... e.g. /sites/company-share/Freigegebene Dokumente/Work"
|
|
|
|
# Check if pathQuery contains search terms (words without proper path structure)
|
|
validPathPrefixes = ['/sites/', '/Documents', '/documents', '/Shared Documents', '/shared documents']
|
|
if not any(pathQuery.startswith(prefix) for prefix in validPathPrefixes):
|
|
return False, f"Invalid pathQuery '{pathQuery}'. This appears to be search terms, not a valid SharePoint path. Use findDocumentPath action first to search for folders, then use the returned folder path as pathQuery."
|
|
|
|
return True, None
|
|
except Exception as e:
|
|
logger.error(f"Error validating pathQuery '{pathQuery}': {str(e)}")
|
|
return False, f"Error validating pathQuery: {str(e)}"
|
|
|
|
def detectFolderType(self, item: Dict[str, Any]) -> bool:
|
|
"""
|
|
Detect if an item is a folder using improved detection logic.
|
|
|
|
Parameters:
|
|
item (Dict[str, Any]): Item from SharePoint API response
|
|
|
|
Returns:
|
|
bool: True if item is a folder, False otherwise
|
|
"""
|
|
try:
|
|
# Use improved folder detection logic
|
|
if 'folder' in item:
|
|
return True
|
|
|
|
# Try to detect by URL pattern or other indicators
|
|
webUrl = item.get('webUrl', '')
|
|
name = item.get('name', '')
|
|
|
|
# Check if URL has no file extension and looks like a folder path
|
|
if '.' not in name and ('/' in webUrl or '\\' in webUrl):
|
|
return True
|
|
|
|
return False
|
|
except Exception as e:
|
|
logger.error(f"Error detecting folder type: {str(e)}")
|
|
return False
|
|
|