Liz Agent Research: AI Best Practices for Investment Advisors
P3 - LowResearch on AI agent best practices for Liz trading assistant — Top 10 improvements with priorities; key finding: extended thinking + reasoning scaffold are highest-impact quick wins
Overview
Comprehensive research report on AI agent best practices for the Liz trading assistant. Full report at ~/clawd/projects/hedge/docs/LIZ_AGENT_RESEARCH.md (716 lines, 35KB).
Key Findings
Architecture
- Liz uses the correct ReAct (Reason+Act+Observe) pattern — tool-use agentic loop
- Best practice: prefer atomic tools for data retrieval, composite tools for common workflows
- Multi-agent financial systems (e.g., MarketSenseAI) use specialized sub-agents: macro, sector, fundamental, technical, sentiment → synthesis
Top 10 Improvements (Prioritized)
- Extended thinking (LOW effort, VERY HIGH impact) — enable
budget_tokens: 8000-16000for complex analysis - Reasoning scaffold in system prompt — add explicit step-by-step analysis instructions
- Cross-session memory — Liz forgets everything between sessions; add Supabase persistence
- Optimized tool responses — add
summaryfield to every tool response for LLM consumption - Increase max_tokens (2000 → 8000) + add streaming
- Bias prevention — add explicit uncertainty/bear-case requirements to system prompt
- Proactive alerts — scheduled signal scanning, regime change alerts (HIGH effort)
- Prompt caching — 40-60% cost reduction on static system prompt + tools
- Few-shot examples — 2-3 example exchanges in system prompt
- Composite tool bundles — get_trade_setup(), get_morning_briefing(), get_401k_status()
Data Sources Research (JPMorgan ICAIF 2024)
- Fundamental data (earnings, revenue, cash flow) = most important
- Sentiment SCORES > full news text (equal performance, fewer tokens, less bias)
- Omitting news entirely often improves performance by reducing recency bias
Financial Data Formatting for LLMs (Daloopa 2025)
- LLMs lose table structure when data is flattened to text
- Use hierarchical JSON with preserved relationships
- Include human-readable interpretations alongside raw numbers
- Always include
summaryfield with 1-2 sentence interpretation
Prompt Engineering for Finance
- Chain-of-Thought prompting dramatically improves financial analysis accuracy
- Analogy-Driven CoT (AD-FCoT) grounds reasoning in historical precedents
- FinCoT (2025) injects domain-specific reasoning blueprints at each step
- Explicit bias prevention: always require bear case, uncertainty levels, invalidation conditions
Current Liz Gaps
- No cross-session memory (biggest UX gap)
- max_tokens = 2000 too low for complex analysis
- No reasoning scaffold in system prompt
- No explicit bias prevention instructions
- Tool responses not optimized for LLM consumption (raw API JSON)
- No proactive alerting
- No prompt caching (paying full token cost per call)
Sources
25+ sources including: Anthropic building-effective-agents, OpenAI agent SDK, JPMorgan ICAIF 2024, MarketSenseAI 2025, FinCoT 2025, AD-FCoT 2025, Daloopa financial data guide, AWS financial AI patterns, RAGflow 2025 review, ArXiv LLM investment bias paper, McKinsey agentic evaluation.
Created: Fri, Feb 27, 2026, 9:01 PM by bob
Updated: Fri, Feb 27, 2026, 9:01 PM
Last accessed: Wed, Apr 1, 2026, 11:53 PM
ID: a2da2520-c73f-4f63-9c9f-b804181a97de