The must-have resource for anyone who wants to experiment with and build on the OpenAI Vision API. This repository serves as a hub for innovative experiments, showcasing a variety of applications ranging from simple image classifications to advanced zero-shot learning models. It's a space for both beginners and experts to explore the capabilities of the Vision API, share their findings, and collaborate on pushing the boundaries of visual AI.
Experimenting with the OpenAI API requires an API 🔑. You can get one here.
- 100 API requests per single API key per day.
- Can't be used for object detection or image segmentation. We can solve this problem by combining GPT-4V with foundational models like GroundingDINO or Segment Anything (SAM). Please take a look at the example and read our blog post.
webcamgpt.mov
- Set-of-Mark Prompting Unleashes Extraordinary Visual Grounding in GPT-4V by Jianwei Yang, Hao Zhang, Feng Li, Xueyan Zou, Chunyuan Li, Jianfeng Gao
- The Dawn of LMMs: Preliminary Explorations with GPT-4V(ision) by Zhengyuan Yang, Linjie Li, Kevin Lin, Jianfeng Wang, Chung-Ching Lin, Zicheng Liu, Lijuan Wang
- GPT-4 System Card by OpenAI
- How CLIP and GPT-4V Compare for Classification
- Experiments with GPT-4V for Object Detection
- Distilling GPT-4 for Classification with an API
- DINO-GPT4-V: Use GPT-4V in a Two-Stage Detection Model
- First Impressions with GPT-4V(ision)
We would love your help in making this repository even better! Whether you want to add a new experiment or have any suggestions for improvement, feel free to open an issue or pull request.
If you are up to the task and want to add a new experiment, please look at our contribution guide. There you can find all the information you need.