URL: https://arxiv.org/pdf/2504.06771v1
Ai, Help Me Think—but For Myself: Assisting People In Complex Decision-Making By Providing Different Kinds Of Cognitive Support
Can you give me a very clear explanation of the core assertions, implications, and mechanics elucidated in this paper?
Okay, let's break down the core assertions, implications, and mechanics of the paper "AI, Help Me Think—but for Myself".
Core Problem Addressed:
- Traditional AI decision support often relies on providing recommendations (like "buy this stock").
- This recommendation-centric approach faces challenges:
- Users may rely inappropriately (too much or too little) on the AI.
- It can lead to user deskilling.
- It's often difficult for users to integrate AI recommendations into their own complex reasoning process, especially when factors outside the AI's model ("unobservables") are important.
- It forces users to reason backward from a proposed solution rather than forward towards their own.
Proposed Alternative & Core Assertion:
- The paper proposes and investigates an alternative AI interaction model called ExtendAI.
- Assertion 1: Instead of giving a direct recommendation, ExtendAI works by taking the user's own written rationale for a decision and providing feedback embedded within that rationale. The goal is to augment and refine the user's thinking process as they reason forward.
- Assertion 2: This ExtendAI approach, compared to a traditional recommendation approach (termed RecommendAI), will lead to different effects on the user's decision-making process, their perception of the AI, and the quality of their decisions.
Mechanics (How it Works & How it Was Studied):
- Two AI Systems Compared:
- RecommendAI: Functions like a typical recommender. Based on the user's current portfolio and profile, it suggests specific actions (e.g., buy/sell specific ETFs). The user receives this list and then decides.
- ExtendAI: Requires the user to first write down their reasoning (their rationale) for intended actions (e.g., "I want to buy ETF X because of trend Y, and sell ETF Z because of risk Q"). The AI (an LLM) then processes this rationale and returns the user's text with embedded feedback (in bold), pointing out potential issues, diversification concerns, overlooked factors, etc., based on the user's own stated logic. It was specifically designed not to give direct ETF recommendations.
- Study Design:
- Use Case: Financial investment decisions using Exchange-Traded Funds (ETFs). Chosen for its complexity and open-ended nature.
- Task: Participants (N=21, with investing experience) used a simulated trading platform to build and revise an ETF portfolio over three time steps (2024, 2026, 2028) based on a provided investor profile and synthetic market news/data.
- Procedure: Within-subjects design.
- Step 1 (2024): Familiarisation, no AI.
- Step 2 (2026): Use either RecommendAI or ExtendAI (randomized order).
- Step 3 (2028): Use the other AI system.
- Data Collection: Mixed methods - questionnaires (confidence, satisfaction, cognitive load, trust, etc.), interaction logs, portfolio analysis (diversification metrics), and qualitative exit interviews.
- AI Implementation: Both systems used GPT-4o, prompted with ETF data, portfolio status, and the investor profile. ExtendAI's prompt additionally included the user's rationale and instructions to provide embedded, non-recommendation feedback.
Key Findings & Assertions Supported/Nuanced:
- ExtendAI:
- Integrated better with users' own thinking and decision process.
- Led to slightly better decision outcomes (e.g., improved portfolio diversification).
- Encouraged more reflection and user engagement.
- Resulted in perceptions of being better informed (confidence aligned better with satisfaction).
- Required more cognitive effort (due to writing the rationale).
- Was sometimes perceived as less insightful or novel because its feedback was tied to the user's initial thoughts.
- RecommendAI:
- Was faster and required less cognitive effort upfront.
- Was better at providing novel insights or suggesting new directions ("out of the box").
- Led to higher initial confidence but lower satisfaction after seeing outcomes (suggesting potential miscalibration or less ownership).
- Was preferred by about half the users, often for its convenience and actionability.
- Some found its recommendations harder to verify ("black box").
- Overall: User preference was split evenly, indicating different users value different aspects (control/engagement vs. ease/novelty).
Core Implications:
- Viable Alternative: ExtendAI demonstrates a feasible and potentially effective alternative to recommendation-centric AI, especially for complex tasks where user reasoning and agency are valued.
- Design Tensions: Designing AI for decision support involves navigating inherent tensions:
- Actionability vs. Cognitive Engagement: How specific should advice be? (Easy to act on vs. encouraging thought).
- Novelty vs. Consistency: How much should AI challenge vs. align with user thinking? (Valuable new ideas vs. easier integration).
- Timing: When is the best moment to introduce AI input? (Not too early to anchor, not too late to be ignored).
- Integration Focus: The ExtendAI approach shifts the focus from just calibrating trust/reliance on a final recommendation towards how AI can be integrated into the human reasoning process itself. This integration might make evaluation easier for the user.
- User Agency Matters: Many users in complex domains desire to remain central to the decision-making process and value tools that support their thinking rather than replace it, even at the cost of higher effort.
- Natural Language Interface: Using natural language for rationale input (ExtendAI) forces reflection but adds friction. Future designs could explore ways to reduce this friction while retaining the benefits.
In essence, the paper argues that AI can support complex decision-making not just by giving answers (RecommendAI), but by engaging with and enhancing the user's own reasoning process (ExtendAI), leading to different trade-offs in user experience, effort, and potentially decision quality.
Can you explain the value of this in basic terms? Like you're talking to a CEO. So what? What's the bottom line here?
Okay, let's cut to the chase.
Right now, AI often just gives recommendations – "buy this," "approve that." It's fast, but for complex decisions, your people might blindly follow it, ignore it completely, or struggle to fit the advice into their own understanding of the situation. This can lead to mistakes, low adoption, or people feeling sidelined.
We tested a different approach: an AI that acts more like a thinking partner. Instead of just giving the answer, it takes your employee's own plan and reasoning, and gives targeted feedback on their thinking. It helps them refine their own strategy.
So what? What's the bottom line?
- Better Complex Decisions: For high-stakes decisions where nuance and human judgment matter (like finance, strategy, complex diagnostics), this "thinking partner" AI (we called it ExtendAI) helped people make slightly better, more robust decisions because it integrated with their expertise, rather than trying to replace it. They understood the 'why' behind the final choice.
- Improved AI Integration & Trust: People felt more in control and engaged with this partner AI. They were more likely to actually use the AI's input effectively because it built on their own work, rather than feeling like a "black box" recommendation (which was our RecommendAI). This means smoother adoption into critical workflows.
- Reduced Risk: By encouraging reflection, this approach can reduce the risk of your team over-relying on potentially flawed AI recommendations in complex scenarios. They are thinking with the AI, not just taking orders from it.
- It's Not One-Size-Fits-All: Some people preferred the quick recommendations, others the thinking partner. The key takeaway is that for different tasks and different people, you might need different kinds of AI support. Simply deploying recommendation bots everywhere isn't the optimal strategy.
In short: For your most critical, complex decisions, designing AI to enhance your team's thinking process, rather than just giving answers, can lead to better outcomes, better adoption, and lower risk. It's about augmenting intelligence, not just automating commands.
Can you give me an analogy or metaphor that will help explain this to a broad audience?
Okay, here are a couple of analogies to explain the difference between the two AI approaches:
Analogy 1: The GPS vs. The Expert Co-Pilot
- RecommendAI is like a standard GPS: You plug in your destination, and it just tells you, "Turn left here," "Turn right there." It gives you the final instructions. You follow them, hopefully getting where you need to go, but you might not understand why that's the best route or what alternatives exist. You're passively receiving directions.
- ExtendAI is like having an expert co-pilot sitting next to you: You first say, "Okay, I'm planning to take this highway, then cut across on this road because I think it's faster." The co-pilot looks at your plan on the map and says, "That highway is good, but that crossroad often gets jammed around this time. Have you considered this slightly longer route that avoids the main traffic light? Also, remember you wanted to stop for gas, and there's a better station along this other way." They aren't just giving you a route; they're engaging with your reasoning and helping you refine your own plan based on your goals and potential blind spots.
Analogy 2: The Ghostwriter vs. The Editor
- RecommendAI is like hiring a ghostwriter: You tell them roughly what you want an article about, and they deliver a finished piece. You can use it, maybe tweak it, but the core thinking and structure came from them.
- ExtendAI is like working with a good editor: You write the first draft of your article, putting down your ideas and arguments. The editor then goes through your text and adds comments and suggestions directly within it (like using "Track Changes"): "This point needs more evidence." "Have you considered the opposing viewpoint here?" "This paragraph could be clearer if you rephrased it like this." They help you strengthen your own writing and clarify your own thoughts, making the final piece better while keeping it fundamentally yours.
The Bottom Line: Both approaches aim to help, but RecommendAI gives you a finished answer to evaluate, while ExtendAI engages with your process of getting to an answer, helping you think more effectively for yourself along the way.
Can you provide a specific example that best illustrates the key points elucidated in this paper?
Okay, let's use a specific example based on the paper's investment scenario, but simplified to highlight the core differences.
Imagine David, who needs to decide how to invest an additional $10,000 into his existing ETF portfolio. His goal, like in the study, is long-term growth with moderate risk, focusing on diversification.
Scenario 1: Using RecommendAI (The Direct Advisor)
- David's Action: He opens the AI tool and clicks "Get Recommendation."
- RecommendAI Output: The AI analyzes his current holdings (let's say he has a US S&P 500 ETF and a Global Tech ETF) and the market data. It outputs:
- "Recommendation:
- Buy $5,000 of VGK (Vanguard FTSE Europe ETF): Adds European diversification.
- Buy $3,000 of ICLN (iShares Global Clean Energy ETF): Taps into a growth sector, aligns with moderate risk tolerance.
- Buy $2,000 more of your S&P 500 ETF: Increases core holding stability."
- "Recommendation:
- David's Experience & Decision Process (Illustrating Key Points):
- Low Initial Effort: David didn't have to articulate his own strategy first.
- Novelty: The AI suggested specific ETFs (VGK, ICLN) he might not have considered immediately.
- Actionability: The advice is very direct and easy to execute.
- Potential Issues (from the paper):
- Integration Difficulty: David now has to evaluate these specific recommendations. Why VGK and not another Europe ETF? Is ICLN really suitable now, given recent news (which the AI might ignore)? He has to reason backward from the AI's solution.
- Cognitive Engagement: He might just accept the recommendations without deep thought ("The AI knows best") or reject them if they feel off, potentially missing the underlying diversification logic.
- Ownership/Trust: If the investments do poorly, he might blame the AI. If they do well, he might feel less ownership. Verifying the AI's reasoning is harder ("black box").
Scenario 2: Using ExtendAI (The Thinking Partner)
- David's Action: He first thinks about his strategy and types his rationale into the AI tool:
- "My current portfolio is heavy on US tech. I want to diversify geographically and maybe add a bit more risk for growth. I'm thinking of adding an Emerging Markets ETF for growth ($5k) and maybe a Healthcare ETF ($5k) for stability, as it seems less volatile than tech."
- ExtendAI Output: The AI processes David's rationale and returns his text with embedded feedback (bolded):
- "My current portfolio is heavy on US tech. I want to diversify geographically and maybe add a bit more risk for growth. Good goal. Diversifying away from US tech concentration is wise. I'm thinking of adding an Emerging Markets ETF for growth ($5k) Emerging markets (like EEM) offer growth potential but can be volatile, aligning with adding risk. Ensure you're comfortable with that volatility level. Also, consider potential overlap with your existing Global Tech ETF, as some tech companies operate heavily in emerging markets. and maybe a Healthcare ETF ($5k) for stability, as it seems less volatile than tech. Healthcare (like IXJ) can add stability. However, many broad healthcare ETFs are also heavily US-focused. To truly boost geographic diversification, have you considered a European-focused ETF (like VGK) instead, or perhaps a global ex-US healthcare option if available? This might balance your US exposure better."
- David's Experience & Decision Process (Illustrating Key Points):
- Higher Initial Effort: David had to articulate his plan first.
- Refinement, Not Just Novelty: The AI didn't necessarily suggest brand new ideas but refined his existing ones (pointing out volatility, overlap, US-focus in healthcare).
- Better Integration: The feedback is directly tied to his reasoning. He can easily see why the AI is suggesting alternatives (e.g., VGK addresses his stated goal of geographic diversification better than a US-heavy healthcare ETF).
- Higher Cognitive Engagement: David is prompted to think more deeply about his choices ("Is EEM too volatile for me?", "Good point about US focus in IXJ").
- Ownership/Trust: The final decision feels more like his, informed by the AI partner. Evaluating the AI's feedback is easier because it's contextualized within his own thoughts.
In a Nutshell:
RecommendAIgave David a ready-made answer, requiring less upfront thought but potentially more effort to evaluate and integrate trustfully.ExtendAIforced David to think first, then helped him improve his own thinking, leading to potentially better understanding and integration, even if it felt like more work initially.
This example shows the trade-offs: speed/novelty/actionability (RecommendAI) versus reflection/integration/ownership/refined reasoning (ExtendAI).