A Lightweight AI Engine for Skin Cancer Detection on Wearable Devices

Skin cancer is one of the most common cancers globally – and early detection significantly improves the chances of successful treatment. Unfortunately, many people lack access to dermatologists or advanced diagnostic tools. This research addresses the problem by bringing AI-based diagnostics to low-cost wearable devices. What did the authors do? Used MobileNetV2: A compact neural network architecture optimized for mobile environments. With transfer learning, the model was fine-tuned to classify skin lesions as cancerous or non-cancerous. ...

July 24, 2025

SOPHIA: Enhancing Slow‑Thinking in Large Vision‑Language Models

In recent years, Large Vision‑Language Models (LVLMs) have shown impressive abilities to understand and generate text about images—but they often struggle with long, multi‑step reasoning. The paper “SOPHIA: Semi‑Off‑Policy Reinforcement Learning for Slow‑Thinking in LVLMs” presents a new approach that significantly improves their capacity for slow‑thinking reasoning. What Is Slow‑Thinking? Slow‑thinking is a deliberate, step‑by‑step reasoning process where the model: Breaks down complex problems into smaller steps, Verifies intermediate conclusions, Provides transparency into each decision. This contrasts with fast, intuitive “snap” judgments and helps avoid hallucinations—invented details not supported by the image. ...

July 23, 2025

The Role of AI in Managing Satellite Constellations

Modern satellite mega-constellations—groups of hundreds or thousands of small satellites working together—are transforming how we connect the world. Yet, managing these networks presents unique challenges: constantly moving nodes, limited onboard computing power, and a need to minimize communication delays. The ConstellAI project, supported by the European Space Agency, explores how artificial intelligence (AI) can optimize two critical tasks: Data Routing: Choosing the best path through the network to send data quickly and reliably. Resource Allocation: Distributing limited resources (bandwidth, power, time slots) among satellites and ground stations. Data Routing with Reinforcement Learning Traditional routing algorithms, like finding the shortest path on a map, don’t account for traffic jams (long queues) at network nodes. ConstellAI uses a technique called reinforcement learning (RL). In RL, a software agent learns from experience: it tries different routes, observes delays, and gradually discovers which paths minimize overall transit time. ...

July 22, 2025

On the Fundamental Limitations of Dual Static CVaR Decompositions in Markov Decision Processes

When making decisions—from financial investments to routing autonomous vehicles—we care not only about average outcomes but also about risk. A widely used risk metric is the Conditional Value at Risk, or CVaR, defined for confidence level $\alpha\in(0,1)$ by: $$ CVaR_\alpha(X) =\inf_{\xi}{\xi + \tfrac{1}{1-\alpha},E[(X-\xi)_+]}. $$ In their recent paper, Godbout and Durand (2025) examine how to reliably compute this metric in Markov Decision Processes (MDPs). They reveal that the most common method—the dual decomposition—suffers from inherent limitations. ...

July 21, 2025

PinFM: Foundation Model for User Activity Sequences at a Billion-Scale Visual Discovery Platform

The paper “PinFM: Foundation Model for User Activity Sequences at a Billion-Scale Visual Discovery Platform” introduces a $>$20B-parameter transformer pretrained on Pinterest user interaction sequences. Its goal is to build a universal sequence model applicable to various recommendation tasks, including content ranking, related Pins, and personalized feeds. Background and Motivation Traditional recommendation systems rely on specialized models for each task. The explosion of data volume and signal diversity calls for a generalized pretraining–finetuning paradigm. PinFM was developed to: ...

July 20, 2025

GradNetOT: Learning Optimal Transport Maps with GradNets

Optimal Transport (OT) is the mathematical problem of moving “mass” from one distribution to another in the most efficient way. Think of reshaping a pile of sand into a new shape with minimal effort. GradNetOT is a novel machine‑learning method that learns exactly these efficient maps using neural networks equipped with a built‑in “bias” toward physically correct solutions. What Is Optimal Transport? Classic formulation: Given two probability distributions (e.g., piles of sand and holes to fill), find a mapping that moves mass at minimal total cost. Monge’s theorem: For certain costs (like squared distance), the optimal map is the gradient of a convex function satisfying a Monge–Ampère equation. The GradNetOT Approach GradNetOT leverages a special neural network architecture called a Monotone Gradient Network (mGradNet) to represent convex functions implicitly. By enforcing convexity and monotonicity, the network’s output gradient automatically yields a valid OT map. ...

July 19, 2025

Unstable Power: How Sharpness Drives Deep Network Learning

The paper “Understanding the Evolution of the Neural Tangent Kernel at the Edge of Stability” by Kaiqi Jiang, Jeremy Cohen, and Yuanzhi Li explores how the Neural Tangent Kernel (NTK) evolves during deep network training, especially under the Edge of Stability (EoS) regime. What is the NTK? The Neural Tangent Kernel (NTK) is a matrix that captures how tiny weight changes affect network outputs on each training example. It lets us analyze neural networks with tools from kernel methods, offering theoretical insights into learning dynamics. What is the Edge of Stability? When training with a large learning rate $\eta$, the largest eigenvalue of the NTK (or the loss Hessian) exceeds the stability threshold $2/\eta$ and then oscillates around it. This phenomenon, called Edge of Stability, combines elements of instability with phases of rapid learning. Key Findings Alignment Shift Higher $\eta$ leads to stronger final Kernel Target Alignment (KTA) between the NTK and the label vector $y$. ...

July 18, 2025

RiemannLoRA: A Unified Riemannian Framework for Ambiguity-Free LoRA Optimization

In recent years, Low‑Rank Adaptation (LoRA) has become a cornerstone technique for parameter‑efficient fine‑tuning of large language models (LLMs) and diffusion models. By injecting low‑rank matrices into pre-trained weights, LoRA drastically reduces memory and compute requirements, enabling rapid experimentation and deployment. However, practitioners face two persistent challenges: Initialization ambiguity: Different low‑rank factor pairs $$A, B$$ can represent the same adapted weight update $AB^\top$, leading to unstable or suboptimal starts. Redundant parameterization: Without a canonical representation, gradient updates can wander through equivalent parameter configurations. The RiemannLoRA framework, introduced by Bogachev et al., offers a unifying geometric viewpoint that removes these ambiguities and yields faster, more stable fine‑tuning. ...

July 17, 2025

A Neural Network Model of Complementary Learning Systems: Pattern Separation and Completion for Continual Learning

Standard neural networks often suffer from catastrophic forgetting, where learning new tasks degrades performance on previously learned tasks. In contrast, the human brain integrates new and old memories through two complementary memory systems: the hippocampus and neocortex. 1. Objectives The authors aim to build a model that captures: Pattern separation: distinct encoding of similar experiences, Pattern completion: reconstructing full representations from partial inputs, to support continual learning without loss of previously acquired skills. ...

July 16, 2025

Target Polish: How to Polish Data and Reveal Its True Structure

Imagine you’re analyzing sensor data. Suddenly one sensor shows -999°C. That’s an outlier — a single data point that can completely ruin your analysis. 🧩 What is factorization? Matrix factorization means decomposing data $X$ into two non-negative components: $$ X \approx WH $$ Where $W$ contains “features” and $H$ shows how much of each is needed. 💡 The problem Classical methods like NMF are sensitive to noise and outliers. When data is messy, analysis breaks down. ...

July 15, 2025