Recently Viewed

New

Vision-Language Pre-Training: Basics, Recent Advances, and Future Trends by Zhe Gan 9781638281320

No reviews yet Write a Review
RRP: $165.75
Booksplease Price: $150.93
Booksplease saves you

  Bookmarks: Included free with every order
  Delivery: We ship to over 200 countries from the UK
  Range: Millions of books available
  Reviews: Booksplease rated "Excellent" on Trustpilot

  FREE UK DELIVERY: When You Buy 3 or More Books - Use code: FREEUKDELIVERY in your cart!

SKU:
9781638281320
MPN:
9781638281320
Available from Booksplease!
Availability: Usually dispatched within 5 working days

Frequently Bought Together:

Total: Inc. VAT
Total: Ex. VAT

Description

Humans perceive the world through many channels, such as images viewed by the eyes, or voices heard by the ears. Though any individual channel might be incomplete or noisy, humans can naturally align and fuse information collected from multiple channels in order to grasp the key concepts needed for a better understanding of the world.

One of the core aspirations in Artificial Intelligence (AI) is to develop algorithms that endow computers with an ability to effectively learn from multimodal (or, multi-channel) data. This data is similar to sights and sounds attained from vision and language that help humans make sense of the world around us. For example, computers could mimic this ability by searching the most relevant images to a text query (or vice versa), and by describing the content of an image using natural language. Vision-and-Language (VL), a popular research area that sits at the nexus of Computer Vision and Natural Language Processing (NLP), aims to achieve this goal.

This monograph surveys vision-language pre-training (VLP) methods for multimodal intelligence that have been developed in the last few years. Approaches are grouped into three categories: (i) VLP for image-text tasks, such as image captioning, image-text retrieval, visual question answering, and visual grounding; (ii) VLP for core computer vision tasks, such as (open-set) image classification, object detection, and segmentation; and (iii) VLP for video-text tasks, such as video captioning, video-text retrieval, and video question answering. For each category, a comprehensive review of state-of-the-art methods is presented, and the progress that has been made and challenges still being faced are discussed, using specific systems and models as case studies. In addition, for each category, advanced topics being actively explored in the research community are presented, such as big foundation models, unified modeling, in-context few-shot learning, knowledge, robustness, and computer vision in the wild, to name a few.

Book Information
ISBN 9781638281320
Author Zhe Gan
Format Paperback
Page Count 204
Imprint now publishers Inc
Publisher now publishers Inc
Weight(grams) 294g

Reviews

No reviews yet Write a Review

Booksplease  Reviews


J - United Kingdom

Fast and efficient way to choose and receive books

This is my second experience using Booksplease. Both orders dealt with very quickly and despatched. Now waiting for my next read to drop through the letterbox.

J - United Kingdom

T - United States

Will definitely use again!

Great experience and I have zero concerns. They communicated through the shipping process and if there was any hiccups in it, they let me know. Books arrived in perfect condition as well as being fairly priced. 10/10 recommend. I will definitely shop here again!

T - United States

R - Spain

The shipping was just superior

The shipping was just superior; not even one of the books was in contact with the shipping box -anywhere-, not even a corner or the bottom, so all the books arrived in perfect condition. The international shipping took around 2 weeks, so pretty great too.

R - Spain

J - United Kingdom

Found a hard to get book…

Finding a hard to get book on Booksplease and with it not being an over inflated price was great. Ordering was really easy with updates on despatch. The book was packaged well and in great condition. I will certainly use them again.

J - United Kingdom