I can’t believe I’ve had an original idea in wanting a face system monitor. The face would change, skin color, eye size, ear flapping, mouth screaming, whatever, and the user could configure it how they wanted. Then I could get a quick system overview just by looking at the face in the corner of my screen. I’d code it up myself, but for 2 reasons. I think I’d tire of it quickly, as badly as I want to see it. I don’t even know where to start with coding anything more than some bash scripting.


Yeah, it would be nice to have an animator and a programmer work together for stuff like this.
On the other hand, Blender has Python scripting, meaning a lot of Blender users would be capable of that.
So all that is required for a programmer is to create Python endpoints and then provide a similar interface that would work on pre-rendered graphics.
I think I got the art part settled: https://www.spriters-resource.com/ms_dos/doomdoomii/asset/27876/
If no one else does this, I think I might