Articles

Build Apps with a Click

Build Apps with a Click

This blog post focuses on new features and improvements. For a comprehensive list, including bug fixes, please see the release notes. Introduced app templates for streamlined app creation. We now provide pre-built, ready-to-use templates that expedite the app creation process. Each template comes with a range of resources, such as datasets, models, workflows, and modules, allowing … Read more

Dangers Of AI – Data Exploitation

Dangers Of AI – Data Exploitation

Introduction – Dangers Of AI – Data Exploitation Artificial intelligence profoundly influences sectors ranging from national security to daily life. As neural networks perform increasingly complex tasks, AI’s role in society expands. Yet, this growth brings an array of risks, particularly in the realm of data exploitation. Financial institutions leverage AI for risk assessments, while … Read more

AI Alignment Is Trivial – by Monica Anderson

AI Alignment Is Trivial – by Monica Anderson

A debate concerning AI Alignment is upon us. We hear ridiculous claims about AIs taking over and killing all humans. These claims are rooted in fundamental 20th Century Reductionist misunderstandings about AI. These fears are stoked and fueled by journalists and social media and cause serious concerns among outsiders to the field. It’s time for … Read more

What We Learned from a Year of Building with LLMs (Part III): Strategy – O’Reilly

What We Learned from a Year of Building with LLMs (Part III): Strategy – O’Reilly

We previously shared our insights on the tactics we have honed while operating LLM applications. Tactics are granular: they are the specific actions employed to achieve specific objectives. We also shared our perspective on operations: the higher-level processes in place to support tactical work to achieve objectives. Learn faster. Dig deeper. See farther. But where do … Read more

How distributed training works in Pytorch: distributed data-parallel and mixed-precision training

How distributed training works in Pytorch: distributed data-parallel and mixed-precision training

In this tutorial, we will learn how to use nn.parallel.DistributedDataParallel for training our models in multiple GPUs. We will take a minimal example of training an image classifier and see how we can speed up the training. Let’s start with some imports. import torchimport torchvisionimport torchvision.transforms as transformsimport torch.nn as nnimport torch.nn.functional as Fimport torch.optim … Read more

Asymmetric Certified Robustness via Feature-Convex Neural Networks – The Berkeley Artificial Intelligence Research Blog

Asymmetric Certified Robustness via Feature-Convex Neural Networks – The Berkeley Artificial Intelligence Research Blog


Asymmetric Certified Robustness via Feature-Convex Neural Networks

TLDR: We propose the asymmetric certified robustness problem, which requires certified robustness for only one class and reflects real-world adversarial scenarios. This focused setting allows us to introduce feature-convex classifiers, which produce closed-form and deterministic certified radii on the order of milliseconds.

Asymmetric Certified Robustness via Feature-Convex Neural Networks – The Berkeley Artificial Intelligence Research Blog


Figure 1. Illustration of feature-convex classifiers and their certification for sensitive-class inputs. This architecture composes a Lipschitz-continuous feature map $\varphi$ with a learned convex function $g$. Since $g$ is convex, it is globally underapproximated by its tangent plane at $\varphi(x)$, yielding certified norm balls in the feature space. Lipschitzness of $\varphi$ then yields appropriately scaled certificates in the original input space.

Despite their widespread usage, deep learning classifiers are acutely vulnerable to adversarial examples: small, human-imperceptible image perturbations that fool machine learning models into misclassifying the modified input. This weakness severely undermines the reliability of safety-critical processes that incorporate machine learning. Many empirical defenses against adversarial perturbations have been proposed—often only to be later defeated by stronger attack strategies. We therefore focus on certifiably robust classifiers, which provide a mathematical guarantee that their prediction will remain constant for an $\ell_p$-norm ball around an input.

Conventional certified robustness methods incur a range of drawbacks, including nondeterminism, slow execution, poor scaling, and certification against only one attack norm. We argue that these issues can be addressed by refining the certified robustness problem to be more aligned with practical adversarial settings.

Read more

Group-equivariant neural networks with escnn

Group-equivariant neural networks with escnn

Today, we resume our exploration of group equivariance. This is the third post in the series. The first was a high-level introduction: what this is all about; how equivariance is operationalized; and why it is of relevance to many deep-learning applications. The second sought to concretize the key ideas by developing a group-equivariant CNN from … Read more

How R U OK? leverages Isentia insights to enhance communications campaigns 

How R U OK? leverages Isentia insights to enhance communications campaigns 

object(WP_Post)#7549 (24) { [“ID”]=> int(35056) [“post_author”]=> string(2) “36” [“post_date”]=> string(19) “2024-10-10 00:11:13” [“post_date_gmt”]=> string(19) “2024-10-10 00:11:13” [“post_content”]=> string(6835) ” The role –and nature– of news and journalism is constantly evolving, from how it is consumed to which voices are trusted. New platforms, the rise of citizen journalists, and shifting news consumption habits are continuing to … Read more

Marek Rosa – dev blog: Introducing Charlie Mnemonic: The First Personal Assistant with Long-Term Memory

Marek Rosa – dev blog: Introducing Charlie Mnemonic: The First Personal Assistant with Long-Term Memory

As part of our research efforts in continual learning, we are open-sourcing Charlie Mnemonic, the first personal assistant (LLM agent) equipped with Long-Term Memory (LTM).  At first glance, Charlie might resemble existing LLM agents like ChatGPT, Claude, and Gemini. However, its distinctive feature is the implementation of LTM, enabling it to learn from every interaction. … Read more