The June '24 AI, Machine Learning, and Computer Vision Meetup covered topics such as leveraging pre-trained text2image diffusion models for zero-shot video editing, improved visual grounding through self-consistent explanations, and combining Hugging Face Transformer models with image data using FiftyOne. Bariscan Kurtkaya discussed the potential of using pre-trained text-to-image diffusion models for video editing without fine-tuning. Dr. Paola Cascante-Bonilla presented her work on enhancing vision-and-language models' ability to localize objects in images by fine-tuning them for self-consistent visual explanations. Jacob Marks demonstrated how the seamless integration between Hugging Face and FiftyOne simplifies connecting datasets and models, enabling more effective data-model co-development. The next Meetup is scheduled for July 3rd, featuring talks on performance optimization for multimodal LLMs, five handy ways to use embeddings in AI, and responsible and unbiased genAI for computer vision.