II. Precautions
- Non-Medical
- Not for Medical Care
- The author uses several software packages and programming languages to develop FPnotebook content
- For convenience, a dozen pages in FPNotebook are dedicated to quick notes on content creation
- Quality and Safe Usage
- Generative AI is imperfect and requires human oversight, evaluation and iterative improvement
- Remember: LLMs generate answers by stringing together the statistically most likely next word
- The answer may not be correct, complete, legal, ethical or responsible
- Always operate with guard rails and oversight
- Integrating AI as a learning resource in medicine
- Conditions
- At work or school, limit AI tools to officially approved agents
- Maintaining patient privacy (e.g. HIPAA) and patient safety are paramount
- Always verify AI output with reputable sources
- Journal clubs
- Compare the accuracy with standard journal articles and medical texts
- Critique the AI responses
- Review the output of an AI tool analyzing the quality of a journal article
- Review accuracy of AI extracted statistics from an article (e.g. NNT, Likelihood Ratio)
- Practice patient interactions
- Motivational Interviewing with a Chat-Bot
- Difficult conversation practice (e.g. Breaking Bad News, Difficult Clinical Encounter)
- Mock patient interviews (e.g. guess the medical condition)
- Documentation review
- Have an AI agent review a de-identified SOAP encounter note for clarity and substance
- Other learning strategies
- Use AI tools to generate a flash card system (chat gives question, user gives answer)
- Log effective prompts for later reuse
- References
- (2026, Jan) Presc Insights, No. 420112
- Rincón (2025) BMC Med Educ 25(1):526 +PMID: 40221725 [PubMed]
- Conditions
III. Types: LLM Models
-
General
- Text data passed to LLMs is tokenized
- Words are broken into common segments (~4 chars each)
- Numbers are broken into groups of 3 numbers each
- Large frontier models typically require payment in the form of price per million tokens
- Cost is based on the combination of input tokens, compute tokens and output tokens
- Text data passed to LLMs is tokenized
- Large Frontier Models (run in the cloud, 100s of billions to trillions or parameters)
- OpenAI GPT
- Anthropic Claude
- Google Gemini
- Deepseek
- Grok
- Openai/gpt-oss-120b
- Smaller Models (may be run locally with Ollama, model variants with<10-20 Billion parameters)
- Meta LLama
- Mistral
- Microsoft Phi-3
- Alibaba Qwen
- Google Gemma
- Deepseek
- Openai/gpt-oss-20b
- Resources
- Vellum Leaderboard
IV. Types: LLM Usage Tiers
- Level 1: Transformers (multi-shot prompting)
- Chat with a model via Model Specific Web Interface (e.g. Chat GPT)
- Chat with a model via Consolidator (e.g. Perplexity)
- Chat with a model via API (e.g. python in Jupyter Notebook)
- Create json with an array of dictionary objects (contains entire history of chat)
- role: system, user, assistant, tool, developer
- content: message
- Create json with an array of dictionary objects (contains entire history of chat)
- Level 2: Agents (one-shot prompting)
- Give a model extensive system instructions to be used with every query
- Give the agent access to a personal or business knowledge base in various forms
- Data may be vectorized for rapid access and context (Retrieval-Augmented Generation or RAG)
- Agent masters domain and task specific queries
- Level 3: Agentic AI Workflows (Virtual Assistants)
- Agents are used in various combinations with one another to form an overall workflow
- Agents may be used in serial or parallel, with data passed along the chain of agents
- Workflow may include if-then branching logic
- Workflows may be autonomous, but typically involve human oversight
- Interfaces
- Web Interfaces (e.g. Cassidy, TypingMind)
- Coded Interfaces (e.g. python)
- Agents are used in various combinations with one another to form an overall workflow
V. Techniques: Prompting Pearls
- Prompting instructions to ensure quality Generative AI answers
- Stop and Think deeply about the question
- Think Step by Step
- If you do not know the answer, say so (do NOT hallucinate)
- Recursive use of AI Transformers and Agents
- Prompt: "What questions should I answer so that you can generate the best answer to my question?"
- Prompt: "I asked another model the question X, and they responded with answer Y. Evaluate their answer."
VI. Technique: System Instructions for Agent Development
- Have AI chat generate the system instructions for you
- Give it the agents general tasks to reach target output (ultimate goal for the agent)
- Ask the AI chat to ask you 5-10 questions related to the agent development it needs to start
- Answer the questions, and any follow-up questions
- Have the AI chat generate a draft of the system instructions into fields as shown below
- System instruction fields (in Markdown for ChatGPT or XML for Claude)
- Role
- Assistant primary function, expertise and personality
- Context
- Background information (business info, common scenarios, company values)
- Instructions
- Step-by-step guides and protocols
- Address general and specific scenarios
- Criteria
- Benchmarks on which to grade the quality of the output
- Identify strict rules (Do's and Don'ts)
- Examples
- Sample scenarios (at least 2-3) with their expected outputs
- Role
- References
- Tyler Fisk, Maven Course: Agentic AI Foundations
VII. Resources: Useful Commercial AI Medical Applications
- Open Evidence
- PM Cardio (Queen of Hearts)
VIII. Resources: Useful General AI Tools
- Online Search (run multiple LLMs)
- Typing Mind
- Perplexity
- Frontier LLM
- Chat-GPT
- Google Gemini
- Anthropic Claude
- Local LLM Use
- Ollama (LLM)
- Pinokio (image and video generation)
- Hugging Face
IX. Resources: Learning
- AI Engineer Core Track: LLM Engineering, RAG, QLoRA, Agents
- AI Engineer Agentic Track: The Complete Agent & MCP Course