- 2024/10/17 AirSim Vs Gazebo Comparison
- 2024/10/15 Airsim Ubuntu 22.04 Setup Guide
- 2024/10/02 AirSim End-to-End Learning
- 2024/09/30 7 Lessons from AirSim
- 2024/08/29 TartanAir: AirSim simulation dataset for simultaneous localization and mapping (SLAM)
- 2023/12/06 End-to-End Deep learning for Autonomous Driving based on AirSim
- Post-Quantum Cryptography (PQC)
- Quantum Key Distribution (QKD)
- 2024/12/31 2025 Expert Quantum Predictions — PQC And Quantum Cybersecurity
- 2024/12/10 PQCrypto Usage & Deployment
- 2024/09/04 An in-depth look at the NIST PQC algorithms
- 2024/09/02 Go 1.23 PQC Quantum-Safe TLS
- 2024/10/09 LLaVA-Llama: An All-Around Text Processor
- 2024/09/18 Multimodal RAG: Chat with Videos and the Future of AI Interaction
- 2024/09/05 Introducing LLaVA V1.5 7B on GroqCloud - Qwen2-Vision, MiniCPM are better
- 2024/07/30 LLaVA Multimodel Image Search
- 2024/07/17 LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models
- 2024/06/18 LLaVA: Large Language and Vision Assistant
Florence-2 is a lightweight vision-language model open-sourced by Microsoft under the MIT license. The model demonstrates strong zero-shot and fine-tuning capabilities across tasks such as captioning, object detection, grounding, and segmentation.
Florence-2, released by Microsoft in June 2024, is a foundation vision-language model. This model is very attractive because of its small size (0.2B and 0.7B) and strong performance on a variety of computer vision and vision-language tasks.
- 2024/07/26 Florence-2: Microsoft's Latest Vision-Language Model
- 2024/07/17 Launch: Deploy Florence-2 with Roboflow
- 2023/04/26 Google Chrome Dino Bot using Image Recognition | Python
- 2023/03/23 Automate Chrome Dino Game using Python
- 2022/08/06 Using Q learning to play the Chrome Dinosaur game
- 2021/08/24 How to play Google Chrome Dino game using reinforcement learning
- 2020/11/08 Chrome Dinosaur Game using Python – Free Code Available
- 2020/10/23 How to Hack the Hidden Google Chrome Dinosaur Game
- 2024/07/10 PaliGemma: A versatile 3B VLM for transfer - SigLIP-So400m vision encoder and the Gemma-2B language model
- 2024/06/27 Welcome Gemma 2 - Google’s new open LLM 🤗
- 2024/06/27 Gemma 2 is now available to researchers and developers
- 2024/06/16 Let’s play with PaliGemma!
- 2024/06/05 Key Challenges in Current Vision Language Models (VLMs)
- 2024/06/04 YOLOv10: The Dual-Head OG of YOLO Series
- 2024/05/10 Video: Robo erectus – W1 quadruped robot evolves to stand and walk upright
- 2024/03/30 This four-legged robot learned parkour to better navigate obstacles
- 2024/03/27 ANYmal can do parkour and walk across rubble
- 2024/03/14 Autonomous ANYmal Quadruped Robot Performs Parkour
- 2024/01/04 Spot Robot : The Design details and working
- 2023/10/31 Small Robotic Dog Takes Giant Parkour Leaps
- 2024/05/18 How to Train a KAN Model on the Titanic Dataset for Kaggle
- 2024/05/15 What are the key concepts of Kolmogorov Arnold networks?
- 2024/05/15 Kolmogorov-Arnold Networks: A Comprehensive Guide to Neural Network Advancement
- 2024/05/11 Kolmogorov–Arnold Networks (KAN) Are About To Change The AI World Forever
- 2024/05/10 [Kolmogorov-Arnold Networks (KANs) Might Change AI As We Know It, Forever](https://www.linkedin.com/pulse/kolmogorov-arnold-networks-kans-might-change-ai-w
- 2024/02/19 Mask2Former: Semantic, Panoptic, and Instance Segmentation with One Architecture
- 2023/12/06 Introducing Scale’s Automotive Foundation Model
- 2023/11/07 YOLOP v2 Explained: A Deep Dive into Panoptic Driving Perception
- 2023/10/06 Make your own self-driving RC car using Python and Raspberry Pi (Part 1 - Introduction)
- 2023/08/30 YOLOPv2: Better, Faster, Stronger for Panoptic Driving Perception
- 2023/08/28 What is Mask2Former? The Ultimate Guide