Daniel B. Chapman

Member
  • Content count

    30
  • Joined

  • Last visited

Community Reputation

4 Neutral

About Daniel B. Chapman

  • Rank
    Greenhorn

Personal Information

  • Occupation
    Scenery, Lighting Design & Production
  • Homepage
    http://www.danielbchapman.com
  • Location
    IL
  1. Gaspar, I would push for a technical answer from the Vision team. I know for a fact if they're routing through CoreVideo (I don't think they are) the eGPU option via Thunderbolt isn't an option as the MacOS won't use the hardware. It should, however, use the OpenGL hardware. In my tests QLab didn't utilize the external hardware (CoreVideo) and Processing did (OpenGL). That was using a Mac Mini 2014 as a test-case which was pretty solid. For whatever it is worth installing the external hardware in an enclosure was a bit of work and less reliable than I liked. My experience with Qt (which is what Vision and Vectorworks are built on) is that it runs very well on Windows, Linux, or Mac so you're probably hardware bound. My recent experience leads me to believe that building a simple PC with a high-end graphics card is less expensive than trying to integrate the external enclosure. You're looking at ~ $200 base for the enclosure which is very similar to the cost of a motherboard and CPU for a PC. That being said, if you're using that card to power other things on your Mac it might be worth it. Best of luck!
  2. Mark, I'm actively using another product for previsualization right now and the NDI workflow is extremely easy. I'm using Open Broadcast Studio to screen-capture media server output and send it over the network so there's actually no hardware in this pipeline. NDI has solid support in Open Frameworks and several capture cards (Blackmagic) have utilities for reading directly into a stream. It is easier than dealing with HDCP issues between devices and is simple to use. I'm able to easily stream two SVGA resolution outputs over wifi with no hardware. If you opt to not support it at a minimum it would be nice to see support for Windows Media Foundation compatible cards and AV Foundation compatible cards on MacOS. If you supported the core frameworks I suspect we would all have a much easier time pulling in video.
  3. Thanks Edward, On my Windows 10 box there's no change for video, I might make it to my studio to fire up the Mac Pro and see if there's any difference. I'll be using the demo as I'm officially "expired" until this bug is fixed but I'm happy to test builds as they come out. I'll reinstall the capture card today and take another look sometime this weekend. I've messed around with the median luminescence ('M' and 'N' are the hotkeys if I remember right) and I can't really get a black level. I end up dropping the contrast to make it reasonable but you end up with that weird blown out color as a result. I'm happy to hear video is on the list. As a programmer who does a lot of 3D and 2D graphics I bet it isn't that big an issue once you get a solid test-box together. I really, really, really recommend you guys use a clean box for testing. As I let you know my ancient backup computer does still show Spout/Virtual Cams in Windows 10x64 and I bet it is either the age of the hardware or something funky since that was upgraded from Windows 7. Also, don't forget about my other post where I show the green/magenta bug in the Mac webcam. I think that's probably a solid clue.
  4. Mark, I think the general problem with the way Vision is attempting to handle video is that you are not providing solid hardware support and the only reliable support seems to be CITP which is pretty similar but far more complex to implement for most applications because it manages a lot more than video. NDI's SDK is here. I have not personally coded up an implementation (though I'm probably adding it to my media servers after this opera). It is really just a UDP video stream that you can tap into. https://www.newtek.com/ndi/sdk/ Essentially the pipeline is just: [NDI Stream Name]::[Frame #][some headers with sizing][pixels] Since that's pushed over IP it is available on local host or on your network. It has a couple little preview image things that get pushed but in general you just pick a name and then decode it on whatever device you want. Here's a link to two utilities I use because I'm using OpenGL calls most of the time: Windows I just take a Spout broadcast and throw it over the network http://spout.zeal.co/download-spout-to-ndi/ On Mac I run http://techlife.sg/TCPSyphon/ which is basically the same thing. This takes the video hardware entirely out of the chain because you're taking a texture and mapping that to a video stream that is decoded on the other end. It really is pretty simple, especially if you are (and I assume you are) just taking a texture and mapping that to your light source for video. As to your concerns about converters: yes, that's a problem but it is significantly less of a problem than trying to support all the native capture cards out there (and it is clear your native hooks do not work at the moment). The IP solution bypasses this because there are a variety of utilities to get the format into NDI. There's a "pro version" here for Blackmagic cards: https://itunes.apple.com/us/app/ndi-source/id1097524095?ls=1&mt=12 which makes some pretty cheap USB3/Thunderbolt hardware available or you. Frankly, I'd rather see Syphon/Spout connectivity, but I operate at a pretty high technical level for video/projection design as I write my own software for most of it. I think having some simple connectors available would be a very quick way to get around this problem. I suspect NDI is going to be a standard that sticks around for quite a while. I first used it on a major corporate event and I'm seeing it everywhere now. At the end of the day I don't actually care how I get a video stream into Vision I just need to be able to get a video stream into Vision. NDI makes multiple inputs a lot easier and in many cases you can completely bypass a capture card with software. That being said, running a HDMI cable out of a Mac into the back of my PC wouldn't really bother me if it meant I could render projection mapping effects easily for a director. In many ways I think supporting Blackmagic devices would be a good choice as they are relatively low cost (Intensity Shuttle) and the SDK is available for Windows/Mac so you don't have to do the hard work or depend on the operating system. At the end of the day I just want to get a video source into Vision. 2.3 just doesn't cut it for professional renderings (and I did make that work with an AverMedia PCIe card).
  5. Dominic, As of this point, no, I don't think it is fixed. If you get it working I would absolutely love to hear what your system is so I can track down what's happening on mine. I just tried 2018 on a Windows 10 and had the same, result. I've had some luck with Vision 2.3 and an Aver Media card but the rendering quality just isn't there for my clients. I'm being a pest here on the forums in the hopes that this will elevate the issue to "needs to be fixed" rather than "backlogged". Here's the card I'm using with Vision 2.3: https://www.amazon.com/AVerMedia-Definition-Composite-Component-C027/dp/B002SQE1O0/ref=sr_1_1?ie=UTF8&qid=1505419715&sr=8-1&keywords=avermedia+capture+card+pcie Edward has been very helpful with trying to make this work but I think the bottleneck is on the developer side of this problem. It is frustrating because Vision is definitely the easiest workflow right now but 2.3 definitely has issues with black-levels and ambient lighting to get something "dark" working. I've attached a screenshot so you know what you're getting into. These are the same models (and no, I didn't pick that ugly orange color, that's an R02 I think). As a designer trying to figure out if my shots work or doing focus points 2.3 is fantastic. If I'm trying to show this to a client it is a no-go. If this is a road block for you I'd encourage you to contact support. I think I'm the only one bugging them on this issue so it isn't a priority. Best of luck! (The dark one is Vision 2017, the lighter one is 2.3 with 0 ambient lighting)
  6. ** by "not plug anything into" I mean I literally can not purchase a capture card that will work with Vision 2017/2018 on Windows 10 or Mac OS and render video inside Vision. This isn't a case of me trying to skirt a hardware requirement. I would happily pay $300 to get this working right now and the answer is I can not.
  7. Mark, As a developer myself I can understand the hurdles. I'd also like to point out that video input has been broken in Vision 4, Vision 2017, and Vision 2018 with some, but not all, capture cards working in Vision 2.3. I have (4) computers that I can test against (a surface, a MacBook Pro 2016 Touch, a MacBook Pro 2010, and a box I built myself running windows 10 that's nothing fancy) and none of these have working video input. Mac OS ends up the closest with some bizarre pixelated green/magenta image of my webcam. Considering the fact video is broken in Vision as it stands you guys might want to explore "universal" connectors like NDI as it seems your ability to support hardware is hampered at best. I could be wrong and a video patch is on the way but I downloaded the Vision 2018 demo today and I get the same broken interface with no available inputs for video projection. While I'm on the topic it would be nice to see you integrate: * Blackmagic inputs (the SDK is pretty well documented) https://www.blackmagicdesign.com/support/ Why? Because they are solid affordable capture cards that run on Mac and Windows * Syphon inputs on Mac (again, very well documented software and a simple texture-share given the fact you're rendering in OpenGL this is borderline trivial) http://syphon.v002.info/ Why? Because it is native to Mac OS and used in the VJ pipeline by basically every media server. This bypasses the hardware issue entirely so it seems like an obvious choice. * Spout inputs on Windows (which is basically Syphon for Windows) http://spout.zeal.co/ Why? Because it is Syphon for Windows. It bypasses the hardware chain entirely and is a simple shared texture. * Support for Virtual Webcams in general which would render all the above moot as you could use Syphon or Spout to make virtual inputs OR something like NDI which has converters for Blackmagic, Syphon, Spout, Capture Cards, Web Cams, Other media servers and CITP protocols. Since this runs of TCP/IP it also renders the hardware question moot. As a software company the phrase "renders the hardware question moot" should be of interest to you. I've been fairly patient on this issue but I'm currently evaluating Capture Nexum and Realizzer for purchase in my next project because Vision is broken for video. I've already suspended my subscription until this issue is fixed. And I suspect that will mean I'm running Vision 2017 for a very long time at this rate. "Which media servers are using this protocol I haven't looked at them in a while so I'm a bit behind the times." This isn't really a valid question at this point given the state of projection design. Not everyone is running a Hippotizer or a D3 on gigs. Many of us use Watchout, QLab, Resolume, or a variety of other pipelines (Touch Designer/Isodora) in our work at the regional level for theater and dance and the inability to plug anything into Vision is extraordinarily frustrating.
  8. I've been recently using NDI in my previz workflow to pass streams from machine to machine and it is definitely easier than any of the other solutions I've worked with recently. Can I request that we get this into Vision to replace the clunky video capture problems? https://www.newtek.com/ndi/ Here's a link to the open frameworks plugin which gives some idea of how easy this actually is: https://github.com/leadedge/ofxNDI I think the receiver code (video source) is less than 600 lines of code interfacing with the SDK
  9. In 2017 I'm unable to sort the scene graph to arrange the fixtures in a way that makes sense. Am I missing something? Is there a better way to select fixture to focus them (I have nearly 250 conventional fixtures on this model).
  10. I'd like to bump this topic, would it be possible to get a response to this even if the answer is "we don't have a workflow"? I'm about ready to head into pre-production and I thought this was "likely fixed" in SP4.
  11. I'm going to reference this long standing thread as the baseline for this post: I have a simple project requirement: 1) Run QLab or some other media server on a computer 2) Pipe that media server into Vision (PC) as a Video Input/Camera Input to be visualized for the design team so I can integrate video and lighting in a video rendering. Vision theoretically supports video capture cards but we don't have a list of approved devices. I don't particularly care if this is Vision 2 32-bit, Vision 2 64bit, Vision 4, or Vision 2017, I just need this to work for this (simple) project. Has anyone had luck getting camera to work in the last couple years? I'm open to any suggestion (including piping it in via OpenFrameworks or Processing if I have to).
  12. Kevin, I can't stress how silly this is right now. Virtual Cameras (not Spout/Syphon) (which are cameras to the operating system) should be a fairly trivial to get working and would allow me--and I suspect others--to do 90% of what we need to do for single-source designs. By way of continuing the discussion here's where we are with basic hardware on the operating system (photos to follow): As you can see with Windows none of the streams are detected. I made of point of using XSplit broadcaster rather than spout to prove a point. Skype is happily running in the corner an detects this as a webcam (I should note that Vision detected Spout in the past using this method). We have a little better luck here with OSX as the FaceTime camera is finally listed but you can see the quality of that image is useless--and rather interesting. I would think that the built in hardware would qualify as the basic check and it isn't functional. If we have a piece of hardware that works in a test environment can we list it? I'd like to get this working ASAP and I'm fine with quirks. Lighting Designs often incorporate video at this point in time and I would like to use Vision to visualize my designs. If that isn't/can't happen then I need to find a new tool that isn't projecting low-resolution video at a cardboard model in my basement. (For whatever it is worth I was expecting the OSX to fail, my other Macbook Pro (2010) didn't list the camera the last time I tried nor did the 2012 iMac at the office. This is on a new Macbook Pro (2016/Touch) so maybe there was better luck on the driver.)
  13. I'd love a response on this thread. It has been roughly six months since I re-raised this issue and I'm curious why there's so little response especially considering the broken state of Vision 2017 in regards to video capture. Here's a link to a simple OpenFrameworks example (I'm going out on a limb to say you're using Qt so it isn't really all that foreign as OpenFrameworks is just a collection of standard OpenGL libraries). The receiver code is less than 500 lines in total. If we can't get resonable capture card support I would like the ability to implement it myself via Spout or Syphon. https://github.com/leadedge/Spout2/tree/master/SpoutSDK/Example/ofSpoutExample/src/ofSpoutReceiver
  14. Edward, Wouldn't it make more sense to implement a shared texture capture rather than try to support every card under the sun? The VJ community has had wild success with Syphon on Mac and a large number of media servers support Spout/Syphon or both (for the few cross platform devices). The lack of video capture in Vision has effectively killed my workflow. Vision 2017 does not support a FaceTime camera out of the box (Windows or OSX) or any other "Virtual" cameras that I've tried to creatively pipe data in. I have no idea what you're using on the back end but it would seem to me that you need to implement a shared texture. Can we please get an update on this? I now have a bug report out that's coming close to a year on this issue. We're officially in "unacceptable" territory here for customer support--it is clear that I'm not the only designer who is annoyed that Vision is useless for visualizing projection designs at the moment.
  15. Can we get a list of the supported video capture cards? I would really like to see Blackmagic and Virtual Webcams added to the list. This has been a very rough year for projector previsualization in Vision and I'd like to see if this is solved/working.