template management for workflow automation

This commit is contained in:
ValueOn AG 2025-11-04 21:14:03 +01:00
parent 78157324ac
commit 94128bfc46
6 changed files with 207 additions and 393 deletions

View file

@ -1,285 +0,0 @@
# Pydantic Class Enhancement Proposal
## Format Tracking & Validation Alignment
**Date:** 2025-11-02
**Purpose:** Align validation logic with prompt requirements, enable workflow-level validation, and track expected file formats
**Simplified Approach:** Use existing document metadata (name, size, format, mimeType) - no summary fields needed
---
## Executive Summary
This proposal addresses:
1. **Validation alignment**: What prompts ask for matches what validators check
2. **Workflow-level validation**: Check ALL deliverables from ALL tasks against original user request
3. **Format tracking**: Track expected formats (list) at workflow and task levels
4. **Adaptive task planning**: Next task uses ALL workflow data (messages, document metadata) to refine objective
**Key Simplification:** Actions deliver documents with metadata (as today). No summary fields needed - use existing document metadata.
---
## 1. ActionResult Class Changes
**File:** `gateway/modules/datamodels/datamodelChat.py` (lines 483-521)
### NO CHANGES NEEDED
**Current Structure (KEEP ALL - ALL USED):**
- ✅ `success: bool` - Used by validation
- ✅ `error: Optional[str]` - Used for error handling
- ✅ `documents: List[ActionDocument]` - Contains document metadata (name, data, mimeType)
- ✅ `resultLabel: Optional[str]` - Used for document routing
**Documents already provide all needed metadata:**
- `documentName` - File name
- `documentData` - Content
- `mimeType` - MIME type (can derive format from this)
**No summary field needed** - document metadata is sufficient.
---
## 2. TaskResult Class Changes
**File:** `gateway/modules/datamodels/datamodelChat.py` (lines 718-736)
### NO CHANGES NEEDED
**Current Structure (KEEP ALL - ALL USED):**
- ✅ `taskId: str` - Task identification
- ✅ `status: TaskStatus` - Task status tracking
- ✅ `success: bool` - Success flag
- ✅ `feedback: Optional[str]` - Task feedback
- ✅ `error: Optional[str]` - Error message
**Document metadata available from workflow:**
- Can extract delivered formats from documents in workflow messages
- No need to store separately - use existing document metadata
---
## 3. TaskStep Class Changes
**File:** `gateway/modules/datamodels/datamodelChat.py` (lines 790-825)
### Modify
- Change `expectedFormat: Optional[str]``expectedFormats: Optional[List[str]]`
- Keep `dataType` and `qualityRequirements` as-is
### Modified Class:
```python
class TaskStep(BaseModel):
id: str
objective: str
dependencies: Optional[list[str]] = Field(default_factory=list)
successCriteria: Optional[list[str]] = Field(default_factory=list)
estimatedComplexity: Optional[str] = None
userMessage: Optional[str] = Field(
None, description="User-friendly message in user's language"
)
# Format details extracted from intent analysis
dataType: Optional[str] = Field(
None, description="Expected data type (text, numbers, documents, etc.)"
)
expectedFormats: Optional[List[str]] = Field(
None, description="Expected output file format extensions (e.g., ['docx', 'pdf', 'xlsx']). Use actual file extensions, not conceptual terms."
)
qualityRequirements: Optional[Dict[str, Any]] = Field(
None, description="Quality requirements and constraints"
)
```
### Register Labels
Update:
```python
"expectedFormats": {"en": "Expected Formats", "fr": "Formats attendus"}
```
---
## 4. ChatWorkflow Class Changes (for Workflow-Level Tracking)
**File:** `gateway/modules/datamodels/datamodelChat.py` (find ChatWorkflow class)
### Add (if not exists)
```python
expectedFormats: Optional[List[str]] = Field(
None,
description="List of expected file format extensions from user request (e.g., ['xlsx', 'pdf']). Extracted during intent analysis."
)
```
Note: `_workflowIntent` is already stored as a dict (not a model field), so `expectedFormats` can be extracted from there, but having it as an explicit field makes it easier to query.
---
## 5. ActionItem Class Review
**File:** `gateway/modules/datamodels/datamodelChat.py` (lines 652-715)
### Current Structure (ALL USED - KEEP):
- ✅ `id: str` - Used for action identification
- ✅ `execMethod: str` - Used for action execution
- ✅ `execAction: str` - Used for action execution
- ✅ `execParameters: Dict[str, Any]` - Used for action execution
- ✅ `execResultLabel: Optional[str]` - Used for document routing
- ✅ `expectedDocumentFormats: Optional[List[Dict[str, str]]]` - Used by action planning
- ✅ `userMessage: Optional[str]` - Used for user communication
- ✅ `status: TaskStatus` - Used for tracking
- ✅ `error: Optional[str]` - Used for error handling
- ✅ `retryCount: int` - Used for retry logic
- ✅ `retryMax: int` - Used for retry logic
- ✅ `processingTime: Optional[float]` - Used for performance tracking
- ✅ `timestamp: float` - Used for ordering/auditing
- ✅ `result: Optional[str]` - Used to store action result text
**NO CHANGES NEEDED** - All attributes are used
---
## 6. Summary of Changes
### Classes to Modify:
1. ✅ **TaskStep** - Change `expectedFormat` (str) → `expectedFormats` (List[str])
2. ✅ **ChatWorkflow** - Add `expectedFormats` (optional, for explicit tracking)
### Classes to Review (NO CHANGES):
- ✅ **ActionResult** - Keep as-is, documents already have metadata
- ✅ **TaskResult** - Keep as-is, no summary needed
- ✅ **ActionDocument** - Already correct (documentName, documentData, mimeType)
- ✅ **ActionItem** - All attributes used
- ✅ **Observation** - Already has contentValidation field
- ✅ **TaskItem** - Used for database storage, separate from TaskStep
---
## 7. Implementation Impact
### Files That Will Need Updates:
1. **datamodelChat.py** - Class definitions (this proposal)
- Change `expectedFormat``expectedFormats` in TaskStep
- Add `expectedFormats` to ChatWorkflow (optional)
2. **taskPlanner.py** - Populate `expectedFormats` list instead of single `expectedFormat`
- **Adaptive planning:** Use ALL workflow data (messages, document metadata) to refine next task objective
- Extract delivered formats from workflow documents
- Compare what was delivered vs. what was planned
3. **contentValidator.py** - Use `expectedFormats` list for validation
- **Action-level validation:** Check action results against task objective (already exists)
- **Task-level validation:** Validate THIS task's deliverables against THIS task's expectations
- Uses document metadata (name, size, format, mimeType) - no summaries needed
4. **intentAnalyzer.py** - Fix prompt to ask for actual file format extensions
- Change from conceptual terms ("raw_data", "formatted") to actual extensions ("pdf", "docx", "xlsx")
5. **promptGenerationTaskplan.py** - Ask for `expectedFormats` in task planning
- **Adaptive planning:** Include ALL workflow data (messages, document names/sizes/formats/metadata) when planning next task
- Show what was actually delivered to help refine objective
6. **workflowManager.py** - Pass ALL workflow data to next task planning
- Messages (text content)
- Document metadata (names, sizes, formats, mimeTypes)
- Validation results
### Key Implementation Points:
- **No summary fields:** Use existing document metadata (name, size, format, mimeType)
- **Adaptive task planning:** Next task receives ALL workflow data (messages + document metadata) to refine objective
- **Validation scope:** Task validation checks ONLY that task's actions, not all workflow actions
- **After each action:** Validate against task objective → decide if complete or next action needed
---
## 8. Validation Logic Alignment
### Action-Level Validation (Within Task):
- **When:** After each action execution within a task
- **Checks:** Action results against task objective
- **Against:** Action documents (name, size, format, mimeType metadata)
- **Purpose:** Decide if task is complete or next action needed
- **Triggers:** Continue to next action if incomplete, complete task if done
### Task Planning (Adaptive - Uses ALL Workflow Data):
- **Input:** ALL workflow data available:
- All messages (text content)
- All document metadata (names, sizes, formats/extensions, mimeTypes)
- Previous task validation results
- **Process:**
- Extract delivered formats from all workflow documents
- Compare what was ACTUALLY delivered vs. what was PLANNED
- Refine next task objective:
- Deliver MORE if previous tasks delivered less than expected
- Deliver LESS if previous tasks already delivered more
- Adapt to actual workflow progress
### Task-Level Validation (Task Completion):
- **When:** After ALL actions in a task complete
- **Checks:** Task objective, task `expectedFormats`, task `successCriteria`
- **Against:** Documents from THIS task only (extract formats from document metadata)
- **Purpose:** Verify THIS task delivered what was expected for THIS task scope
- **Output:** Validation result (used in workflow data for next task planning)
### Workflow-Level Validation (Final):
- **When:** After ALL tasks complete
- **Checks:** Original user request, workflow `expectedFormats`, workflow success criteria
- **Against:** ALL documents from ALL tasks (extract formats from document metadata)
- **Purpose:** Final verification that complete workflow delivered what user requested
- **Triggers:** New compensatory task if validation fails (missing deliverables)
---
## 9. Next Steps
1. **Review and approve this proposal**
2. **Implement class changes** in datamodelChat.py
3. **Update intent analyzer prompt** to request actual file format extensions
4. **Update task planning prompt** to request `expectedFormats` list
5. **Update AI generation prompts** to include summary instruction
6. **Implement aggregation logic** for summaries at task/workflow levels
7. **Implement workflow-level validation** method
8. **Update all references** from `expectedFormat` to `expectedFormats`
---
## Questions Answered
**Document metadata:** Use existing document fields (name, size, format from mimeType/extensions) - no summaries needed
**Format extraction:** Extract formats from document metadata (mimeType or file extensions)
**Task validation scope:** Task validation checks ONLY actions in that task, not all workflow actions
**Adaptive planning:** Next task uses ALL workflow data (messages + document metadata) to refine objective
**After each action:** Validate against task objective → decide complete or next action needed
---
## 10. Validation Flow Clarification
### Simplified Flow:
1. **Within Task (Action-by-Action):**
- Action executes → delivers documents with metadata
- Validate action results against task objective
- If incomplete → next action needed
- If complete → task done
2. **Task Planning (Adaptive):**
- Receives: ALL workflow data (messages, document metadata from all previous tasks)
- Extracts: Delivered formats from document metadata (file extensions/mimeTypes)
- Compares: What was actually delivered vs. what was planned
- Refines: Next task objective (may need more/less based on actual progress)
3. **Task Completion:**
- Validate: THIS task's documents (extract formats from metadata) against THIS task's expectations
- Result: Used in workflow data for next task planning
4. **Workflow Completion:**
- Final validation: All documents (extract formats from metadata) meet original user request
- If missing: Create compensatory task
---
**Status:** Ready for implementation after approval

View file

@ -1219,7 +1219,7 @@ class ChatObjects:
def _computeAutomationStatus(self, automation: Dict[str, Any]) -> str:
"""Compute status field based on eventId presence"""
eventId = automation.get("eventId")
return "active" if eventId else "inactive"
return "Running" if eventId else "Idle"
def getAllAutomationDefinitions(self, pagination: Optional[PaginationParams] = None) -> Union[List[Dict[str, Any]], PaginatedResult]:
"""

View file

@ -147,39 +147,39 @@ class ChatService:
else:
# Direct label reference - can be round1_task2_action3_contextinfo format or simple label
# Search for messages with matching documentsLabel to find the actual documents
matchingMessages = []
for message in workflow.messages:
# Validate message belongs to this workflow
msgWorkflowId = getattr(message, 'workflowId', None)
if not msgWorkflowId or msgWorkflowId != workflowId:
if msgWorkflowId:
logger.debug(f"Skipping message {message.id} with workflowId {msgWorkflowId} (expected {workflowId})")
matchingMessages = []
for message in workflow.messages:
# Validate message belongs to this workflow
msgWorkflowId = getattr(message, 'workflowId', None)
if not msgWorkflowId or msgWorkflowId != workflowId:
if msgWorkflowId:
logger.debug(f"Skipping message {message.id} with workflowId {msgWorkflowId} (expected {workflowId})")
else:
logger.debug(f"Skipping message {message.id} with no workflowId (expected {workflowId})")
continue
msgDocumentsLabel = getattr(message, 'documentsLabel', '')
# Check if this message's documentsLabel matches our reference
if msgDocumentsLabel == docRef:
# Found a matching message, collect it for comparison
matchingMessages.append(message)
# If we found matching messages, take the newest one (highest publishedAt)
if matchingMessages:
# Sort by publishedAt descending (newest first)
matchingMessages.sort(key=lambda msg: getattr(msg, 'publishedAt', 0), reverse=True)
newestMessage = matchingMessages[0]
if newestMessage.documents:
docNames = [doc.fileName for doc in newestMessage.documents if hasattr(doc, 'fileName')]
logger.debug(f"Added {len(newestMessage.documents)} documents from newest message {newestMessage.id}: {docNames}")
allDocuments.extend(newestMessage.documents)
else:
logger.debug(f"No documents found in newest message {newestMessage.id}")
else:
logger.debug(f"Skipping message {message.id} with no workflowId (expected {workflowId})")
continue
msgDocumentsLabel = getattr(message, 'documentsLabel', '')
# Check if this message's documentsLabel matches our reference
if msgDocumentsLabel == docRef:
# Found a matching message, collect it for comparison
matchingMessages.append(message)
# If we found matching messages, take the newest one (highest publishedAt)
if matchingMessages:
# Sort by publishedAt descending (newest first)
matchingMessages.sort(key=lambda msg: getattr(msg, 'publishedAt', 0), reverse=True)
newestMessage = matchingMessages[0]
if newestMessage.documents:
docNames = [doc.fileName for doc in newestMessage.documents if hasattr(doc, 'fileName')]
logger.debug(f"Added {len(newestMessage.documents)} documents from newest message {newestMessage.id}: {docNames}")
allDocuments.extend(newestMessage.documents)
else:
logger.debug(f"No documents found in newest message {newestMessage.id}")
else:
logger.error(f"No messages found with documentsLabel: {docRef}")
raise ValueError(f"Document reference not found: {docRef}")
logger.error(f"No messages found with documentsLabel: {docRef}")
raise ValueError(f"Document reference not found: {docRef}")
logger.debug(f"Resolved {len(allDocuments)} documents from document list: {documentList}")
return allDocuments

View file

@ -931,9 +931,26 @@ class MethodSharepoint(MethodBase):
return ActionResult.isFailure(error=f"Invalid pathQuery '{pathQuery}'. This appears to be search terms, not a valid SharePoint path. Use findDocumentPath action first to search for folders, then use the returned folder path as pathQuery.")
# For pathQuery, we need to discover sites to find the specific one
sites = await self._discoverSharePointSites()
if not sites:
all_sites = await self._discoverSharePointSites()
if not all_sites:
return ActionResult.isFailure(error="No SharePoint sites found or accessible")
# If pathQuery starts with /site:, extract site name and filter
if pathQuery.startswith('/site:'):
# Extract site name from /site:Company Share/... format
site_path_part = pathQuery[6:] # Remove '/site:'
if '/' in site_path_part:
site_name = site_path_part.split('/', 1)[0]
else:
site_name = site_path_part
# Filter sites by name (case-insensitive substring match)
sites = self._filter_sites_by_hint(all_sites, site_name)
if not sites:
return ActionResult.isFailure(error=f"No SharePoint site found matching '{site_name}'")
logger.info(f"Filtered to site(s) matching '{site_name}': {[s['displayName'] for s in sites]}")
else:
sites = all_sites
else:
# Step 3: Both pathObject and pathQuery failed - ERROR, NO FALLBACK
return ActionResult.isFailure(error="No valid upload path provided. Either provide pathObject (from findDocumentPath) or a valid pathQuery with specific site information.")
@ -1094,15 +1111,14 @@ class MethodSharepoint(MethodBase):
"""
GENERAL:
- Purpose: Upload documents to SharePoint. Only to choose this action with a connectionReference
- Input requirements: connectionReference (required); documentList (required); fileNames (required); optional pathObject or pathQuery.
- Input requirements: connectionReference (required); documentList (required); optional pathObject or pathQuery.
- Output format: JSON with upload status and file info.
Parameters:
- connectionReference (str, required): Microsoft connection label.
- pathObject (str, optional): Reference to a previous path result.
- pathQuery (str, optional): Upload target path if no pathObject.
- documentList (list, required): Document reference(s) to upload.
- fileNames (list, required): Output file names.
- documentList (list, required): Document reference(s) to upload. File names are taken from the documents.
"""
try:
connectionReference = parameters.get("connectionReference")
@ -1110,14 +1126,13 @@ class MethodSharepoint(MethodBase):
documentList = parameters.get("documentList")
if isinstance(documentList, str):
documentList = [documentList]
fileNames = parameters.get("fileNames")
pathObject = parameters.get("pathObject")
upload_path = pathQuery
logger.debug(f"Using pathQuery: {pathQuery}")
if not connectionReference or not documentList or not fileNames:
return ActionResult.isFailure(error="Connection reference, document list, and file names are required")
if not connectionReference or not documentList:
return ActionResult.isFailure(error="Connection reference and document list are required")
# If pathObject is provided, extract folder IDs from it
if pathObject:
@ -1296,7 +1311,23 @@ class MethodSharepoint(MethodBase):
upload_site_scope = selected_site
# Use the inner path portion as the actual upload target path
upload_paths = [f"/{parsed['innerPath'].lstrip('/')}"]
# Remove document library name from path (same logic as listDocuments)
inner_path = parsed['innerPath'].lstrip('/')
path_segments = [s for s in inner_path.split('/') if s.strip()]
if len(path_segments) > 1:
# Path has multiple segments - first might be a library name
# Try without first segment (assuming it's a library name)
inner_path = '/'.join(path_segments[1:])
logger.info(f"Removed first path segment (potential library name), path changed from '{parsed['innerPath']}' to '{inner_path}'")
elif len(path_segments) == 1:
# Only one segment - if it's a common library-like name, use empty path (root)
first_segment_lower = path_segments[0].lower()
library_indicators = ['document', 'dokument', 'shared', 'freigegeben', 'library', 'bibliothek']
if any(indicator in first_segment_lower for indicator in library_indicators):
inner_path = ''
logger.info(f"First segment '{path_segments[0]}' appears to be a library name, using root")
upload_paths = [f"/{inner_path}" if inner_path else "/"]
sites = [selected_site]
else:
# When using pathObject, check if upload_path is a folder ID or a path
@ -1311,6 +1342,10 @@ class MethodSharepoint(MethodBase):
# Process each document upload
upload_results = []
# Extract file names from documents
fileNames = [doc.fileName for doc in chatDocuments]
logger.info(f"Using file names from documentList: {fileNames}")
for i, (chatDocument, fileName) in enumerate(zip(chatDocuments, fileNames)):
try:
fileId = chatDocument.fileId

View file

@ -1,66 +0,0 @@
{
"template":
{
"overview": "Automated workflow task",
"tasks": [
{
"id": "Task01",
"title": "Main Task",
"description": "Execute automated workflow",
"objective": "Execute automated workflow",
"actionList": [
{
"execMethod": "ai",
"execAction": "webResearch",
"execParameters": {
"prompt": "{{KEY:webResearchPrompt}}",
"list(url)": ["{{KEY:webResearchUrl}}"]
},
"execResultLabel": "web_research_results"
},
{
"execMethod": "sharepoint",
"execAction": "listDocuments",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"pathQuery": "{{KEY:sharepointFolderNameSource}}"
},
"execResultLabel": "sharepoint_source_path"
},
{
"execMethod": "sharepoint",
"execAction": "readDocuments",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"documentList": ["sharepoint_source_path"],
"pathQuery": "{{KEY:sharepointFolderNameSource}}"
},
"execResultLabel": "sharepoint_source_documents"
},
{
"execMethod": "sharepoint",
"execAction": "uploadDocument",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"documentList": ["sharepoint_source_documents","web_research_results"],
"pathQuery": "{{KEY:sharepointFolderNameDestination}}",
"fileNames": ["report.docx"]
},
"execResultLabel": "sharepoint_upload_documents"
}
]
}
]
}
,
"parameters":
{
"connectionName": "connection:msft:p.motsch@valueon.ch",
"webResearchUrl": "https://www.valueon.ch",
"webResearchPrompt": "Wer arbeitet bei ValueOn AG in der Schweiz und was machen die?",
"PromptSharepointSource": "Fasse die Dokumente in einer Liste zusammen",
"sharepointFolderNameSource": "/site:Company Share/Freigegebene Dokumente/15. Persoenliche Ordner/Patrick Motsch/input",
"sharepointFolderNameDestination": "/site:Company Share/Freigegebene Dokumente/15. Persoenliche Ordner/Patrick Motsch/output",
"PromptDeliverable": "Erstelle mir einen Word Bericht der Webanalyse und der Dokumente im Sharepoint"
}
}

View file

@ -0,0 +1,130 @@
{
"sets": [
{
"template":
{
"overview": "Beispiel Web und Sharepoint",
"tasks": [
{
"id": "Task01",
"title": "Main Task",
"description": "Execute automated workflow",
"objective": "Execute automated workflow",
"actionList": [
{
"execMethod": "ai",
"execAction": "webResearch",
"execParameters": {
"prompt": "{{KEY:webResearchPrompt}}",
"list(url)": ["{{KEY:webResearchUrl}}"]
},
"execResultLabel": "web_research_results"
},
{
"execMethod": "sharepoint",
"execAction": "listDocuments",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"pathQuery": "{{KEY:sharepointFolderNameSource}}"
},
"execResultLabel": "sharepoint_source_path"
},
{
"execMethod": "sharepoint",
"execAction": "readDocuments",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"documentList": ["sharepoint_source_path"],
"pathQuery": "{{KEY:sharepointFolderNameSource}}"
},
"execResultLabel": "sharepoint_source_documents"
},
{
"execMethod": "ai",
"execAction": "process",
"execParameters": {
"aiPrompt": "{{KEY:PromptDeliverable}}",
"documentList": ["sharepoint_source_documents","web_research_results"]
},
"execResultLabel": "sharepoint_report"
},
{
"execMethod": "sharepoint",
"execAction": "uploadDocument",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"documentList": ["sharepoint_report"],
"pathQuery": "{{KEY:sharepointFolderNameDestination}}"
},
"execResultLabel": "sharepoint_upload_documents"
}
]
}
]
},
"parameters":
{
"connectionName": "connection:msft:p.motsch@valueon.ch",
"webResearchUrl": "https://www.valueon.ch",
"webResearchPrompt": "Wer arbeitet bei ValueOn AG in der Schweiz und was machen die?",
"PromptSharepointSource": "Fasse die Dokumente in einer Liste zusammen",
"sharepointFolderNameSource": "/site:Company Share/Freigegebene Dokumente/15. Persoenliche Ordner/Patrick Motsch/input",
"sharepointFolderNameDestination": "/site:Company Share/Freigegebene Dokumente/15. Persoenliche Ordner/Patrick Motsch/output",
"PromptDeliverable": "Erstelle mir einen Bericht der Webanalyse und der Dokumente im Sharepoint als Word Dokument"
}
},
{
"template":
{
"overview": "Immobilienrecherche Zürich",
"tasks": [
{
"id": "Task02",
"title": "Immobilienrecherche Zürich",
"description": "Webrecherche nach Immobilien im Kanton Zürich und Speicherung in Excel",
"objective": "Immobilienrecherche im Kanton Zürich zum Verkauf (5-20 Mio. CHF) und speichere Ergebnisse in Excel-Liste auf SharePoint",
"actionList": [
{
"execMethod": "ai",
"execAction": "webResearch",
"execParameters": {
"prompt": "{{KEY:immobilienResearchPrompt}}",
"list(url)": ["{{KEY:immobilienResearchUrl}}"]
},
"execResultLabel": "immobilien_research_results"
},
{
"execMethod": "ai",
"execAction": "process",
"execParameters": {
"aiPrompt": "{{KEY:excelFormatPrompt}}",
"documentList": ["immobilien_research_results"],
"resultType": "xlsx"
},
"execResultLabel": "immobilien_excel_list"
},
{
"execMethod": "sharepoint",
"execAction": "uploadDocument",
"execParameters": {
"connectionReference": "{{KEY:connectionName}}",
"documentList": ["immobilien_excel_list"],
"pathQuery": "{{KEY:sharepointFolderNameDestination}}"
},
"execResultLabel": "immobilien_upload_result"
}
]
}
]
},
"parameters":
{
"connectionName": "connection:msft:p.motsch@valueon.ch",
"sharepointFolderNameDestination": "/site:Company Share/Freigegebene Dokumente/15. Persoenliche Ordner/Patrick Motsch/output",
"immobilienResearchUrl": ["https://www.homegate.ch", "https://www.immoscout24.ch", "https://www.immowelt.ch"],
"immobilienResearchPrompt": "Suche nach Immobilien zum Verkauf im Kanton Zürich, Schweiz, im Preisbereich von 5-20 Millionen CHF. Sammle Informationen zu: Ort, Preis, Beschreibung, URL zu Bildern, Verkäufer/Kontaktinformationen.",
"excelFormatPrompt": "Erstelle eine Excel-Datei mit den recherchierten Immobilien. Jede Immobilie soll eine Zeile sein mit den folgenden Spalten: Ort, Preis (in CHF), Beschreibung, URL zu Bild, Verkäufer. Verwende die Daten aus der Webrecherche."
}
}
]
}