Picture for Oliver Lemon

Oliver Lemon

Heriot-Watt University

AlanaVLM: A Multimodal Embodied AI Foundation Model for Egocentric Video Understanding

Add code
Jun 19, 2024
Viaarxiv icon

Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language Resamplers

Add code
Apr 21, 2024
Figure 1 for Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language Resamplers
Figure 2 for Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language Resamplers
Figure 3 for Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language Resamplers
Figure 4 for Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language Resamplers
Viaarxiv icon

Socially Pertinent Robots in Gerontological Healthcare

Add code
Apr 11, 2024
Figure 1 for Socially Pertinent Robots in Gerontological Healthcare
Figure 2 for Socially Pertinent Robots in Gerontological Healthcare
Figure 3 for Socially Pertinent Robots in Gerontological Healthcare
Figure 4 for Socially Pertinent Robots in Gerontological Healthcare
Viaarxiv icon

NLP Verification: Towards a General Methodology for Certifying Robustness

Add code
Mar 15, 2024
Figure 1 for NLP Verification: Towards a General Methodology for Certifying Robustness
Figure 2 for NLP Verification: Towards a General Methodology for Certifying Robustness
Figure 3 for NLP Verification: Towards a General Methodology for Certifying Robustness
Figure 4 for NLP Verification: Towards a General Methodology for Certifying Robustness
Viaarxiv icon

Visually Grounded Language Learning: a review of language games, datasets, tasks, and models

Add code
Dec 05, 2023
Figure 1 for Visually Grounded Language Learning: a review of language games, datasets, tasks, and models
Figure 2 for Visually Grounded Language Learning: a review of language games, datasets, tasks, and models
Figure 3 for Visually Grounded Language Learning: a review of language games, datasets, tasks, and models
Figure 4 for Visually Grounded Language Learning: a review of language games, datasets, tasks, and models
Viaarxiv icon

Multitask Multimodal Prompted Training for Interactive Embodied Task Completion

Add code
Nov 07, 2023
Viaarxiv icon

Detecting Agreement in Multi-party Conversational AI

Add code
Nov 06, 2023
Figure 1 for Detecting Agreement in Multi-party Conversational AI
Figure 2 for Detecting Agreement in Multi-party Conversational AI
Figure 3 for Detecting Agreement in Multi-party Conversational AI
Figure 4 for Detecting Agreement in Multi-party Conversational AI
Viaarxiv icon

Detecting agreement in multi-party dialogue: evaluating speaker diarisation versus a procedural baseline to enhance user engagement

Add code
Nov 06, 2023
Viaarxiv icon

FurChat: An Embodied Conversational Agent using LLMs, Combining Open and Closed-Domain Dialogue with Facial Expressions

Add code
Aug 30, 2023
Figure 1 for FurChat: An Embodied Conversational Agent using LLMs, Combining Open and Closed-Domain Dialogue with Facial Expressions
Figure 2 for FurChat: An Embodied Conversational Agent using LLMs, Combining Open and Closed-Domain Dialogue with Facial Expressions
Viaarxiv icon

Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering

Add code
Aug 29, 2023
Figure 1 for Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering
Figure 2 for Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering
Figure 3 for Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering
Figure 4 for Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering
Viaarxiv icon