Discrete Latent Structure in Neural Networks, Vlad Niculae, Caio Corro, Nikita Nangia, Tsvetomila Mihaylova, Andre F. T. Martins (9781638285700) — Readings Books

Become a Readings Member to make your shopping experience even easier. Sign in or sign up for free!

Become a Readings Member. Sign in or sign up for free!

Hello Readings Member! Go to the member centre to view your orders, change your details, or view your lists, or sign out.

Hello Readings Member! Go to the member centre or sign out.

Discrete Latent Structure in Neural Networks
Paperback

Discrete Latent Structure in Neural Networks

$152.99
Sign in or become a Readings Member to add this title to your wishlist.

This title is printed to order. This book may have been self-published. If so, we cannot guarantee the quality of the content. In the main most books will have gone through the editing process however some may not. We therefore suggest that you be aware of this before ordering this book. If in doubt check either the author or publisher’s details as we are unable to accept any returns unless they are faulty. Please contact us if you have any questions.

Machine learning (ML) is often employed to build predictive models for analyzing rich data, such as images, text, or sound. Most such data is governed by underlying structured representations, such as segmentations, hierarchy, or graph structure. It is common for practical ML systems to be structured as pipelines, including off-the-shelf components that produce structured representations of the input, used as features in subsequent steps of the pipeline. On the one hand, such architectures require availability of these components, or of the data to train them. Since the component may not be built with the downstream goal in mind, a disadvantage of pipelines is that they are prone to error propagation. On the other hand, they are transparent: the predicted structures can be directly inspected and used to interpret downstream predictions. In contrast, deep neural networks rival and even outperform pipelines by learning dense, continuous representations of the data, solely driven by the downstream objective.

This monograph is about neural network models that induce discrete latent structure, combining the strengths of both end-to-end and pipeline systems. In doing so, not one specific downstream application in natural language processing nor computer vision is assumed, however the presentation follows an abstract framework that allows to focus on technical aspects related to end-to-end learning with deep neural networks.

The text explores three broad strategies for learning with discrete latent structure: continuous relaxation, surrogate gradients, and probabilistic estimation. The presentation relies on consistent notations for a wide range of models. As such, many new connections between latent structure learning strategies are revealed, showing how most consist of the same small set of fundamental building blocks, but use them differently, leading to substantially different applicability and properties.

Read More
In Shop
Out of stock
Shipping & Delivery

$9.00 standard shipping within Australia
FREE standard shipping within Australia for orders over $100.00
Express & International shipping calculated at checkout

MORE INFO

Stock availability can be subject to change without notice. We recommend calling the shop or contacting our online team to check availability of low stock items. Please see our Shopping Online page for more details.

Format
Paperback
Publisher
now publishers Inc
Country
United States
Date
2 June 2025
Pages
126
ISBN
9781638285700

This title is printed to order. This book may have been self-published. If so, we cannot guarantee the quality of the content. In the main most books will have gone through the editing process however some may not. We therefore suggest that you be aware of this before ordering this book. If in doubt check either the author or publisher’s details as we are unable to accept any returns unless they are faulty. Please contact us if you have any questions.

Machine learning (ML) is often employed to build predictive models for analyzing rich data, such as images, text, or sound. Most such data is governed by underlying structured representations, such as segmentations, hierarchy, or graph structure. It is common for practical ML systems to be structured as pipelines, including off-the-shelf components that produce structured representations of the input, used as features in subsequent steps of the pipeline. On the one hand, such architectures require availability of these components, or of the data to train them. Since the component may not be built with the downstream goal in mind, a disadvantage of pipelines is that they are prone to error propagation. On the other hand, they are transparent: the predicted structures can be directly inspected and used to interpret downstream predictions. In contrast, deep neural networks rival and even outperform pipelines by learning dense, continuous representations of the data, solely driven by the downstream objective.

This monograph is about neural network models that induce discrete latent structure, combining the strengths of both end-to-end and pipeline systems. In doing so, not one specific downstream application in natural language processing nor computer vision is assumed, however the presentation follows an abstract framework that allows to focus on technical aspects related to end-to-end learning with deep neural networks.

The text explores three broad strategies for learning with discrete latent structure: continuous relaxation, surrogate gradients, and probabilistic estimation. The presentation relies on consistent notations for a wide range of models. As such, many new connections between latent structure learning strategies are revealed, showing how most consist of the same small set of fundamental building blocks, but use them differently, leading to substantially different applicability and properties.

Read More
Format
Paperback
Publisher
now publishers Inc
Country
United States
Date
2 June 2025
Pages
126
ISBN
9781638285700