mastouille.fr est l'un des nombreux serveurs Mastodon indépendants que vous pouvez utiliser pour participer au fédiverse.
Mastouille est une instance Mastodon durable, ouverte, et hébergée en France.

Administré par :

Statistiques du serveur :

429
comptes actifs

#ComputerVision

0 message0 participant0 message aujourd’hui

🚀 Excited to share our new paper:

"DynTex: A real-time generative model of dynamic naturalistic luminance textures"

...now published in Journal of Vision!

🔹 Why it matters: Dynamic textures (e.g., fire, water, foliage) are everywhere, but modeling them in real-time has been a challenge. DynTex bridges this gap with a biologically inspired, efficient approach.

🔹 Key innovation: A generative model that captures the spatiotemporal statistics of natural scenes while running in real-time.

🔹 Applications: Computer vision, neuroscience, VR/AR, and more.📖

Read it here: doi.org/10.1167/jov.25.11.2

with Andrew Meso, Nikos Gekas, Jonathan Vacher, Pascal Mamassian and Guillaume Masson

More on: laurentperrinet.github.io/publ

Our lab member Samia Mohinta @Mohinta2892 is featured in this outreach video prepared by the Accelerate Program for Scientific Discovery, explaining in lay terms how we map neural circuits in the fruit fly brain using machine learning approaches, AKA computer vision:
youtube.com/watch?v=68fQ7ZGTxAQ

The eFIB-SEM instrument is featured, setup in the lab, as well as various views of the #MRCLMB.

"In 1966, Seymour Papert proposed the Summer Vision Project, bringing together artificial intelligence researchers for “the construction of a significant part of a visual system” over the course of just a few months. But the goal of solving many problems of computer vision proved overambitious and instead the researchers rediscovered a fact long familiar in vision science: seeing is harder than it looks.

Since the optical studies of Ibn al-Haytham in the 11th century, scientists recognized a gap between the confused mess of visual information hitting the eye and visual experience, where segmented objects appear arrayed in space with clear differences between foreground and background. The assumption for the last millennium has been that there must be unconscious judgments engaged in processing the visual information: segmentation, spatial configuration, object recognition, and so on.

Papert recognized this in part. His optimism stemmed from the idea that the different unconscious judgments necessary for understanding an image could be instantiated in different computer programs. Thus the labor could be divided among different teams, with one team writing a program to detect edges, corners, and other pixel-level information in an image, another forming continous shapes out of these low-level features, a different group arranging the shapes in three-dimensional space, and so on. While the summer project failed, the general approach remained: treat vision not as a single problem, but as a number of discrete subproblems which can be stacked one on top of another...
(...)
Funding for computer vision was often generous because the military was being overwhelmed with spy plane and later satellite images (it was the Cold War, after all). The funding for this approach typically came from either the Department of Defense or the Advanced Research Projects Agency (ARPA, later DARPA), the US military research slush fund."

philippschmitt.com/archive/com

philippschmitt.comA Wiggish History of Computer Vision

Ai2 unveils MolmoAct: Open-source robotics system reasons in 3D and adjusts on the fly - Jiafei Duan, Ai2 researcher, shows MolmoAct controlling a robotic arm. (GeekWire ... - geekwire.com/2025/ai2-unveils- #real-timerobotplanning #alleninstituteforai #computervision #open-sourceai #multimodalai #3dreasoning #seattletech #paulallen #molmoact #robotics #molmo #tech #ai2

🧠 TODAY at #CCN2025 ! Poster A145, 1:30-4:30pm at de Brug & E‑Hall. We've developed a bio-inspired "What-Where" CNN that mimics primate visual pathways - achieving better classification with less computation. Come chat! 🎯

Presented by main author Jean-Nicolas JÉRÉMIE and in cosupervision with Emmanuel Daucé

laurentperrinet.github.io/publ

Our research introduces a novel "What-Where" approach to CNN categorization, inspired by the dual pathways of the primate visual system:

  • The ventral "What" pathway for object recognition

  • The dorsal "Where" pathway for spatial localization

Key innovations:

✅ Bio-inspired selective attention mechanism

✅ Improved classification performance with reduced computational cost

✅ Smart visual sensor that samples only relevant image regions

✅ Likelihood mapping for targeted processing

The results?

Better accuracy while using fewer resources - proving that nature's designs can still teach us valuable lessons about efficient AI.

Come find us this afternoon for great discussions!

Another one of my posts. This one on the topic of AI tools as assistive technology, what's working, what isn't and why, all without the hype that too many people tend to lean into when discussing this technology:

When Independence Meets Uncertainty: My Journey with AI-Powered Vision
A blind user's candid assessment of the promises and pitfalls of current AI accessibility tools
open.substack.com/pub/kaylielf

Kaylie’s Substack · 🤖👁️ From thermostat success to dryer disasters: my honest take on AI vision tools that promise independence but deliver uncertainty. A must-read for anyone curious about the real state of AI accessibility.Par Kaylie L. Fox

"An increasing number of scholars, policymakers and grassroots communities argue that artificial intelligence (AI) research—and computer-vision research in particular—has become the primary source for developing and powering mass surveillance. Yet, the pathways from computer vision to surveillance continue to be contentious. Here we present an empirical account of the nature and extent of the surveillance AI pipeline, showing extensive evidence of the close relationship between the field of computer vision and surveillance. Through an analysis of computer-vision research papers and citing patents, we found that most of these documents enable the targeting of human bodies and body parts. Comparing the 1990s to the 2010s, we observed a fivefold increase in the number of these computer-vision papers linked to downstream surveillance-enabling patents. Additionally, our findings challenge the notion that only a few rogue entities enable surveillance. Rather, we found that the normalization of targeting humans permeates the field. This normalization is especially striking given patterns of obfuscation. We reveal obfuscating language that allows documents to avoid direct mention of targeting humans, for example, by normalizing the referring to of humans as ‘objects’ to be studied without special consideration. Our results indicate the extensive ties between computer-vision research and surveillance."

nature.com/articles/s41586-025

NatureComputer-vision research powers surveillance technology - NatureAn analysis of research papers and citing patents indicates the extensive ties between computer-vision research and surveillance.