shakedown.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A community for live music fans with roots in the jam scene. Shakedown Social is run by a team of volunteers (led by @clifff and @sethadam1) and funded by donations.

Administered by:

Server stats:

264
active users

#computervision

0 posts0 participants0 posts today

Another one of my posts. This one on the topic of AI tools as assistive technology, what's working, what isn't and why, all without the hype that too many people tend to lean into when discussing this technology:

When Independence Meets Uncertainty: My Journey with AI-Powered Vision
A blind user's candid assessment of the promises and pitfalls of current AI accessibility tools
open.substack.com/pub/kaylielf

Kaylie’s Substack · 🤖👁️ From thermostat success to dryer disasters: my honest take on AI vision tools that promise independence but deliver uncertainty. A must-read for anyone curious about the real state of AI accessibility.By Kaylie L. Fox

"An increasing number of scholars, policymakers and grassroots communities argue that artificial intelligence (AI) research—and computer-vision research in particular—has become the primary source for developing and powering mass surveillance. Yet, the pathways from computer vision to surveillance continue to be contentious. Here we present an empirical account of the nature and extent of the surveillance AI pipeline, showing extensive evidence of the close relationship between the field of computer vision and surveillance. Through an analysis of computer-vision research papers and citing patents, we found that most of these documents enable the targeting of human bodies and body parts. Comparing the 1990s to the 2010s, we observed a fivefold increase in the number of these computer-vision papers linked to downstream surveillance-enabling patents. Additionally, our findings challenge the notion that only a few rogue entities enable surveillance. Rather, we found that the normalization of targeting humans permeates the field. This normalization is especially striking given patterns of obfuscation. We reveal obfuscating language that allows documents to avoid direct mention of targeting humans, for example, by normalizing the referring to of humans as ‘objects’ to be studied without special consideration. Our results indicate the extensive ties between computer-vision research and surveillance."

nature.com/articles/s41586-025

NatureComputer-vision research powers surveillance technology - NatureAn analysis of research papers and citing patents indicates the extensive ties between computer-vision research and surveillance.

A friend posted this senior computer vision position on LI. DM me if you want an intro:

linkedin.com/posts/activity-73

From friend: ... fully remote in the US. Deep Sentinel is growing fast, and AI is at the core of our product offerings... role in the full job description: lnkd.in/gZ7iFvnk ... only considering people who have experience building deep learning CV ... unable to work with agencies, offshoring, or sponsor candidates #computervision #machinelearning #hiring

Seeing @neauoire's paper computing explorations, I was reminded of an old idea I proposed at PaperCamp LDN back in 2009: Originally this was vaguely about using origami to create multi-purpose AR markers, but I then extended it to other use cases, some of which could be adapted and be even relevant today, e.g. as a form of proof-of-work, protection against AI crawlers or other form of access control.

Some possible approaches for different use cases:

1) Ask users to perform or series of simple folds and then check results by validating fold lines in the flat sheet paper via computer vision
2) Same as #1, but perform shape recognition/validation of fully folded result
3) Unfold a pre-shared (and pre-folded) origami object to check result by validating fold lines via computer vision
4) Provide instructions for multiple origami creations to create uniquely identifiable objects for computer vision based interactive environments

Number 1-3 are more or less about forms of proving physicality, work, membership. Number 4 is more about using origimi as fiducial markers for general interactions

More ideas/summary from that event:
adactio.com/journal/1546

cc/ @adactio

✍ 📰 𝗨𝗻𝘀𝗲𝗿 𝗝𝗼𝘂𝗿𝗻𝗮𝗹𝗶𝘀𝘁-𝗶𝗻-𝗥𝗲𝘀𝗶𝗱𝗲𝗻𝗰𝗲-𝗣𝗿𝗼𝗴𝗿𝗮𝗺𝗺 𝗴𝗲𝗵𝘁 𝗶𝗻 𝗱𝗶𝗲 𝗻ä𝗰𝗵𝘀𝘁𝗲 𝗥𝘂𝗻𝗱𝗲

👩‍💻 Das Programm bietet Journalist:innen die Möglichkeit, in einem drei- bis sechsmonatigen, bezahlten Aufenthalt in Tübingen zu einem selbst gewählten Thema zu recherchieren.

🤖 Von uns gibt es Einführungen in Themen wie #MachineLearning, #ComputerVision und #Robotik

📧 Bitte bewerben Sie sich mit kurzem Ideenpapier, Anschreiben&Lebenslauf (DE o. ENG) 𝗯𝗶𝘀 𝟮𝟰.𝟬𝟰.𝟮𝟬𝟮𝟱
Bewerbungen bitte per E-Mail an janis.fischer@cyber-valley.de

To avoid a massive OpenCV dependency for a current project I'm involved in, I ended up porting my own homemade, naive optical flow code from 2008 and just released it as a new package. Originally this was written for a gestural UI system for Nokia retail stores (prior to the Microsoft takeover), the package readme contains another short video showing the flow field being utilized to rotate a 3D cube:

thi.ng/pixel-flow

I've also created a small new example project for testing with either webcam or videos:

demo.thi.ng/umbrella/optical-f