- AI + Medicine Newsletter
- Posts
- To fine-tune or not to fine-tune
To fine-tune or not to fine-tune
When adopting general models like GPT-4 and LLaMA-2 to specific areas, such as healthcare, we can use different strategies, including prompt engineering, parameter-efficient fine-tuning, instruction tuning, and re-training from scratch (ordered by technical difficulty to implement, figure 1). People usually believe that fine-tuning with domain-specific knowledge and data can yield better results than prompt engineering. Is it necessary to do fine-tuning, which can cost significant computing resources but probably gain only marginal performance improvement?
Figure 1. LLM adoption techniques (source: Nvidia)
In a recent manuscript published by Microsoft, the authors showed that GPT-4 with a composition of several prompting strategies (named Medprompt, figure 2) can achieve superior performance in multiple medical challenge benchmarks in the absence of fine-tuning. These strategies include (figure 3):
Few-shot prompting: Provide a few examples of correct Q&A. The study developed a kNN-based few-shot, which selects the examples based on K-nearest text embeddings from previous correct Q&A.
Chain-of-thought: Guide the LLM to think step by step
Ensembling: Produce multiple outputs and then consolidate to a consensus output
Another study also found that Flan-T5 with minimal prompt engineering still achieves high fidelity when analyzing clinical notes.
Figure 2. Comparing Medprompt with other models and settings
Figure 3. Medprompt strategy
Boehringer plugs in IBM-trained AI model to boost antibody drug discovery efforts (source)
Boehringer (again) bets $509M on Phenomic’s tumor-targeting tech, as AI biotech plans for busy year (source)
Phenomic (again) enters into strategic research collaboration with Astellas for solid tumor cell therapies (source)
Johnson & Johnson bets heavy on AI-driven drug discovery (source)
Amgen expands pact with Amazon to usher drug manufacturing into the AI era (source)
AstraZeneca will pay Absci up to $247 million to collaborate on the design of an anti-cancer antibody using AI (source).
Other important papers to read
An ethical principle, namely “GREAT PLEA” (Governability, Reliability, Equity, Accountability, Traceability, Privacy, Lawfulness, Empathy, and Autonomy), is proposed for generative AI in healthcare. A Nature comment also recommends that medical professionals, not commercial interests, must drive their development and deployment to protect people’s privacy and safety.
We previously mentioned the need for multimodal AI models for healthcare, and here comes one from the University of Oxford, called Med-MLLM, which supports both image and text data.
Figure 4. Structure of the presented Med-MLLM framework
##########
If you find the newsletter helpful, please consider:
🔊 Sharing the newsletter with other people
👍 Upvoting on Product Hunt
📧 Sending any feedback, suggestions, and questions by directly replying to this email or writing reviews on Product Hunt
🙏 Supporting us with a cup of coffee.
Thanks, and see you next time!
Reply