Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
Hardware Technology

USB/Firewire "Branching" -- Is it Possible? 55

Shaze asks: "I have been doing a lot of overseas web conferencing lately and have been playing around with my two webcams. I was suddenly intrigued by the idea of being able to 'span' or 'branch' both webcams together to create one device pulling signals from both. I did some quick research on the topic to no avail, however many of my colleagues believe this to be possible. Some even going so far as to say they have seen software or drivers that will turn your webcams into one Twain device. Aside from my own meanderings, I can't seem to find any information on anyone or anything even suggesting this is possible. What do you say Slashdot readers: Is this a new application waiting to be marketed or am I just seeing double?"
This discussion has been archived. No new comments can be posted.

USB/Firewire "Branching" -- Is it Possible?

Comments Filter:

  • What do you hope to acomplish by doing this? Stereovision?

    Does it have to be real time?

    In short, why?

    -- MarkusQ

    • I'm guessing he wants two webcams which appear to a piece of video-conferencing software to be a single webcam comprising two images glued together.

      Sounds like he wants a camera aggregator driver which glues together 'n' webcams into a single, larger (or PiP style) image.

      I can't see that this is likely to be supported by anything. Most likely he'll want a piece of software which reads 'n' webcams in realtime and pastes them together (solving any refresh, synchronisation, resolution and format problems) a
      • Most likely he'll want a piece of software which reads 'n' webcams in realtime and pastes them together (solving any refresh, synchronisation, resolution and format problems) and then presents that stream as a new, third aggregated webcam.

        You forgot free, by tomorrow, and with as little work from himself as possible.
    • If his plan was to use three webcams, perhaps he'd be planning to become the next Des e-Arnaz.
    • It barely handles multiple devices at all.

      After all, TWAIN is not SANE [mostang.com].
  • The one thing that jumps into my mind is depth, but this would only work if a mean was taken between both. What about the areas that are seen differently from the cam, there would need to be some filter work done i guess.
  • Uh, what? (Score:1, Redundant)

    by GoRK ( 10018 )
    What is it you're trying to do? The words "Span" and "Branch" mean about as much to me as.. Saying you want the webcams to "Regulate" and "Press"..

    Likewise "One device pulling signals from both" is equally nondescript. I have plugged two USB cameras into one machine before. There is no problem doing this. If you want them to both work at the same time, you'll have to use seperate host controllers or USB 1.1 cameras on a USB 2.0 bus.
    • At least, not under Linux. No shortage of bandwidth on USB 2.0, and the USB software has no trouble handling a squillion cameras all at once.

      Perhaps he's hoping to glue together enough webcams to get decent resolution? (-:
      • Well actually no.. at least not if you use USB 1.1 cameras which most cheap webcams still are. The OS has nothing to do with it.

        USB1.1 devices on a USB2.0 controller can be difficult.. If you didn't put them on seperate controllers you'd have to put them on a USB hub with multiple USB1.1 USB2.0 TT's. The Transaction Translator is the device responsible for translating the USB1.1 data to USB2.0 and vice versa. There is typically only one TT per USB controller in the chipset. Webcams are especially nasty si
  • by jam244 ( 701505 )
    Technically, it is possible to read simultaneously from 2 webcams. It is well within the bandwidth bounds of USB 2.0 and IEEE 1394.

    However, this would probably require a total software rewrite to enable. I don't imagine drivers would need major revisions, fortunately.

    My main question, though, is why? Unless you plan to red/blue shift the objects based on apparent depth and require people to wear stupid glasses, you won't see any interesting effects.

    I think we should be more concerned with bringi
    • Re:Not pointless (Score:4, Informative)

      by Bastian ( 66383 ) on Thursday October 23, 2003 @02:10AM (#7288183)
      At my current job, we use dual FireWire cameras to capture simultaneous images in order to get two concurrent views of moving objects.

      Hooking two cameras to a computer is standard practise in computer vision (not what my current employer does, just another example). It's a really good way to get binocular views of a scene for passive stereo systems.

      Besides, you don't need goofy red/blue glasses to look at stereo pairs of images. Just cross your eyes, for goodness sakes!
      • Please DON'T cross your eyes to view stereograms. This is known as "freeviewing" and I can attest from personal experience that it will eventually damage your vision. Long term use of freeviewing is well known to gradually ruin your vision. So DON'T do it, invest in a cheap plastic stereo viewer, that should only cost you a couple of bucks. Otherwise, it will cost you your vision.
        • exactly how does freeviewing damage your vision? how badly was yours damaged?
          • I had perfect 20/20 vision, and after a few years of freeviewing, I became significantly far-sighted. Forcing yourself to go cross-eyed and deliberately focusing your eyes unnaturally, as is the standard method in freeviewing, will gradually degrade your fine muscle control that allows your eyes to focus accurately.
            • The aging process and gradual enlargement of your lenses makes it more difficult to focus.

              Did this happen starting in your twenties? I find this a more plausible explanation, unless somehow you found occasion to "freeview" at stereograms for hours a day??

              • Yes, I worked at the USGS in my 20s and I spent a LOT of time freeviewing big stereograms. The photogrammetrists all told me not to do it, they all told me about how it would damage my vision, but I did it anyway, and now I regret it. I told them to get me one of those nice Zeiss stereoscopes like the other departments used, but they didn't have the budget (these were big 10in stereograms, and those viewers cost hundreds of bucks).

                You can buy a cheap stereoscope for general use for between $2 and $6, is yo
                • My main point was, how can you be so certain that the normal aging process didn't make you farsighted and make focusing more difficult?
                  • Presbyopia doesn't set in until your 40s, most people with 20/20 vision experience no changes until their 40s. This wasn't normal aging. Sure, aging is part of the process. Do you want to accelerate the process? Or would you rather spend $2 on a stereoscope?
            • I had perfect 20/20 vision, and after a few years of freeviewing, I became significantly far-sighted.

              So since I'm naturally nearsighted, I should do it more so that my vision becomes normal?
              • You can be both nearsighted and farsighted at the same time. That's what bifocals are for. If you don't want to need them, I wouldn't go courting farsightedness just yet.

      • So I'm curious--do you use a Mac for that? If so, are you willing to describe more about what you do/where you work? I know that Macs aren't the only things with FW, but they're most of what I see actually using it...

        You're welcome to reply off-list if you prefer.
      • "At my current job, we use dual FireWire cameras to capture simultaneous images in order to get two concurrent views of moving objects." aka "I do tech support for webcam porn" :D
    • I think we should be more concerned with bringing up the audiovideo quality of videoconferencing.

      Well, that's all well and good, but even though there's one-to-one Internet video conferencing technology good enough that TechWeb thought it was a satellite feed [macdailynews.com] when they saw it last week, it still needs extensions to support one-to-many, many-to-one and many-to-many conferencing.

      And even then you'd still have the onerous task of convincing everyone to use iChat AV [apple.com]. After all, it's an Apple [apple.com] product, a

  • I'm pretty sure I've seen scenarios where a SAN was hooked to two different computers or whatever, which could conceivably happen over FireWire (though Fibrechannel might be more common).

    The catches are:

    1. The thing that's being hooked to both other things being smart enough to realize that it's hooked to two things instead of one, and not get confused.
    2. Actually doing whatever it is you want to do (which starts with knowing what you want to do).

    There are probably computers out there with enough horsepow

    • I've seen old Amigas with toaster cards that can do some interesting stuff with combining two video streams in realtime.
      • Everything the Amiga was doing with the two (actually the toaster has a 4 input switcher; the demo VHS said this OVER and OVER again back in the day... four input switcher! with the four input switcher... the four input switcher...) was being done by the toaster. This is obviously true because even a 68000-based Amiga 2000 can drive the toaster, the Amiga is just an interface. The toaster was far and away a more powerful system than the computer wrapped around it.
    • There are probably computers out there with enough horsepower (stop looking at me like that, you may not borrow the dual G5, my precioussss...) to take 2 typical little video streams from webcams and do something with them in real time.

      Most modern computers today have 3D accelerators in them. Those cards are basically built to do image processing in hardware. (Naturally they are made to transform 3D -> 2D, but the math is largely the same so it doesn't matter.)

      I've only seen stuff about this in academi

    • I once saw a man with a hat that held soda cans and spliced two straws together.

      This guy was able to enjoy 24oz of carbonated refreshment instead of 12, but looked like a total ass.
  • What??? (Score:3, Informative)

    by Stigmata669 ( 517894 ) on Thursday October 23, 2003 @12:53AM (#7287895)
    What I think the writer is asking is if it is possible to have two cameras and stitch the feeds together so it's twice as wide. Like having two monitors but as inputs rather than outputs. I'd assume that it is quite possible to write a program that would listen to both devices and put the feeds next to eachother but I'm not sure why you'd want to do this.
    • Re:What??? (Score:3, Funny)

      by orthogonal ( 588627 )
      What I think the writer is asking is if it is possible to have two cameras and stitch the feeds together so it's twice as wide.

      Yeah! I got an email today telling me they could make it twice as wide!

      Something about a pill.

      Come to think of it, I get one of these emails almost everyday.
    • You can stitch the feeds together to get a panorama, but due to the optical situation with using two cameras, the images will only mesh properly for objects at a certain distance. Stuff that's closer or further away will have "pencil in a glass of water" type discontinuities.
    • When I read the incredibly badly worded and obviously unedited question.
  • I've seen similar (Score:3, Informative)

    by Bastian ( 66383 ) on Thursday October 23, 2003 @02:03AM (#7288165)
    There are a number of computer vision devices composed of a pair of cameras that are treated as a single device. (They must be treated as such in some applications because the cameras must grab images in lockstep.)

    I've never seen something that uses one USB and one Firewire camera, though. To me, it makes just as much sense to use two FireWire cameras or two USB 2.0 cameras since the bandwidth of USB 1.2 is so small. I did some computer vision work as an undergrad with a single Logitech Quickcam Express. The camera grabbed 24 bit 320x244 images at a rather low framerate - something like 10 or 15 fps. This took up 90% of the USB's bandwidth!

    If you want to, you might be able to write a driver that allows you to treat the two as a single TWAIN device, but I imagine the possibility of such a task is highly OS-dependent. I haven't taken a good look to check, but it seems like it would be possible to do so under the Video4Linux framework. MacOS and Windows, I have no idea.
  • The correct softare answer depends on why you want to use two cameras. One solution is to write a "user space driver" that stitches the two image sources into a single picture and creates either a service stream or a virtual device that negotiates the incompatabilities between the cameras. The simplest solution is just to enable your software to use two cameras and two images that are then displayed in two windows or in one data stream or are used to perform stereoscopic vision... or... ah... two angles of
  • If you want stereovision, it's easier to just get a beamsplitting prism rig and attach it to a single camera. These are fairly common in photography shops, here's a good example:

    http://www.3dstereo.com/Merchant2/merchant.mvc? S cr een=PROD&Product_Code=LOR-LICeos&Category_Code=jl- ncam&Product_Count=28

    Another approach that would be more universal (i.e. can put any 2 video signals together) would be to preprocess it with conventional analog TV split screen gadgets, the 2 video inputs are merged
  • by Animats ( 122034 ) on Thursday October 23, 2003 @03:53AM (#7288515) Homepage
    You can read two FireWire cameras simultaneously under Windows 2000. There seems to be a bit of occasional image tearing when two cameras are running. It won't work using the old Video for Windows interface, but the newer Windows API works. Get a pair of Unibrain Fire-i cameras and play with them.

    I have multiple cameras working quite well under QNX, but I wrote that driver. It only uses about 4% of the (1.5GHz) CPU per camera at 640x480 x 15FPS RGB, so it's a modest CPU load. I don't get the image tearing problem Windows does, so the hardware can do the job; it's a Windows problem.

    The Linux FireWire driver subsystem has some major problems in the bus reset area and doesn't handle multiple bus masters properly. The last time I looked at that driver, it didn't allocate isochronous channels properly; everything went on channel 0. That limits you to one active camera. But that may have been fixed by now.

  • Yeah, it's probably possible to do. However, you'd be much better off creating an application that just read from N devices at a time and displayed them in the manner that you want to have them displayed.

    Custom driver == expensive (either in time... or money or both)
  • by Sandman1971 ( 516283 ) on Thursday October 23, 2003 @06:32AM (#7288904) Homepage Journal
    Probably not exactly what you're looking for, but it is another possible solution. I've been using WebcamXP [webcamxp.com] to do just that, but HTML based. It can use up to 5 cameras at the same time (I've used up to 3 at the same time; 2 USB webcams and my camcorder plugged into my Firewire port). I've modified the HTML it generates to put the windows side by side. You'll get the small border around each window, but at least to me, that's not a huge deal.
  • by HughsOnFirst ( 174255 ) on Thursday October 23, 2003 @09:21AM (#7289628)
    The intel Open Source Computer Vision Library [intel.com] They have tools to process stereo usb cameras in real time
  • Where I used to work (Score:3, Informative)

    by phorm ( 591458 ) on Thursday October 23, 2003 @01:19PM (#7292298) Journal
    Wireless camera. TV input card with 2 input channels. Software fast-swapped between the channels and grabbed images.
    • That's very interesting - you mean a regular TV input card that rapidly tuned between 2 input channels coming over the one cable, right? I was thinking of doing something like this for a homebrew PVR. Can you tell me how fast could it swap between channels? (i.e. what fps did you get per channel?)
      • Actually, it was an input-card with 2 input channels on 2 seperate (RCA) jacks if I remember correctly.

        And as for FPS... I think it was a bit jerky as it was used to record between two monitoring devices (where a little stutter is not too bad) and might not be suitable for PVR recording, but then the software (custom-made) wasn't too great either. I'm not sure the channel-swapping was at fault though.
        For hom applications, you could probably get away with a standard TV-input card. The per-channel FPS woul
  • If you just want your conferencing partner to see both your cameras, then why don't you simply launch two copies of the application?
  • lurking with pron tech providers. I think they may seen the commercial opprtunities of this. Why not hunt around where the porn tech providers hang out.

    Haven't they led most of the practial/commercial advances in this area?

HELP!!!! I'm being held prisoner in /usr/games/lib!

Working...