When we need to manipulate something remotely, we use cameras to see, what goes on, and project images from these cameras on the screens.
However, to gain desirable degree of intuitive analysis of the scenery, we need a special sensory device.
First, to imitate walking around the object, we will use a few identical viewing devices providing different views. We will switch between these devices, as needed.
Each such viewing device should support a number of capabilities.
Second, to gain stereo viewing, we will use a pair of cameras looking at the same object, and we will use a type of “gaggles”, where images from different cameras of the pair are presented to different eyes. The distance between centers of the lens of these cameras should be close to the distance between pupils of a person.
Third, to imitate moving of the head to gain better understanding of the scene, we will allow some movement of this pair of cameras relative to the “base” where they are mounted.
Forth, to focus a camera on an object, we will move its lens back-and-force. We will focus both cameras of the assembly simultaneously and we will tilt cameras accordingly to focus both on the same point in space. We will do it in synchrony to avoid causing unpleasant feelings.
Hence, we need an assembly of two cameras placed at a fixed distance. Each camera could be tilted and its lens could be moved to focus on a point in space. This focusing should be done in synchrony for both cameras of the assembly, the way eyes operate. A few such assemblies should be used, that an operator could switch between them to get better view of the area of operations. Images should be passed from an active assembly to “gaggles” providing different views to different eyes.
This assembly should be mounted on a base in the way that it is possible to move its center in all three dimensions and it is possible to turn its “line of view” in both possible dimensions.
We need convenient and intuitive way to control this system, leaving both hands free to do some work.
One mechanism would require placing of radio transmitters on the “gaggles” and radio receivers somewhere in the space, where the operator sits. This mechanism would allow detection of the direction, in which the operator points his face and the place of his head. His “line of view” could be used to set the “line of view” of the active assembly. The position of his head in the space should be used to set position of the two-camera assembly on its base.
In addition, an operator could use feet to control viewing system. With his feet, an operator would move pedals. Some pedals would have springs and thus requiring perpetual application of pressure, as in the case of car’s accelerator, others operating as switches, through discrete kicks on them.
We have three types of operations:
- switching from one assembly to another,
- orienting the line of view of the assembly to point in proper direction,
- focusing the cameras of the assembly at a particular point on the line of view (moving the lens and tilting the cameras).
We should enumerate all two-camera assemblies in some intuitive way and use a kick by a foot to a lever in both directions to switch assemblies according to this enumeration.
Orienting the line of view requires setting of two parameters; placing focus of view on a particular point on this line of view requires setting of the third parameter.
Ideally, it would be nice to have a few different ways of control of viewing for an operator to choose from.
Alexander Liss 4/30/2020