We have seen a new photo mode on cellphones that came from the DSLR camera’s. Portrait mode blurs the bacground for a great effect. Portrait mode selfies is not possible on the S9/S9+ only the S9+ can shoot portrait mode due to its second camera at the REAR.
Only the Iphone X and the Pixel 2/2xl master this feature on the front camera.
Portrait mode video is still not fully possible and requires programs like Adobe After Effects, but with a hack shown in this article, it can come pretty close as well to full after processing of frame by frame editing.
One of the headline features of the dual camera system in recent iPhones is Portrait Mode, an effect that simulates a DSLR-style shallow depth of field by intelligently blurring the background of your photos. Apple took the feature a step further with the iPhone X, adding Portrait Mode selfies and introducing simulated Portrait Lighting.
While these features are currently limited to still photos, future iPhones could someday use the same technology to bring the depth effect to videos, a stunning look currently only possible with high-end video gear or a considerable amount of work in post-production. This feature could be a game changer for videographers, editors, and consumers alike, who have already embraced the iPhone as a serious filmmaking tool. With some effort, the effect is actually possible with today’s iPhones. Here’s how it can be done.
When you take a photo in Portrait Mode on an iPhone today, the depth information associated with the image is stored as a grayscale depth map. iOS uses this depth map to determine which parts of the photo should be blurred and which should remain in focus. This is the same way that 3D artists fake depth of field in 3D renders – the animation software creates a depth map that’s later interpreted by the renderer.
If you want to create the same effect in a video today, there are two common options. The first and easiest is to spend the money on expensive cameras and lenses with a wide aperture – that’s what gives you a shallow depth of field.
A depth map generated by 3D rendering software. Darker areas are in focus and lighter areas are blurred.
The second and more tedious option is to use post-production software like Adobe After Effects to build your own depth maps or video masks in order to specify which parts of a scene should be in focus. This can involve a strenuous process called rotoscoping, a time-consuming task that is often done frame-by-frame.
iOS 11 includes improved developer frameworks that give more access to depth data captured by the iPhone’s cameras. Apple showed off these new capabilities at WWDC 2017 with a sample app called AVCamPhotoFilter. Essentially, this lets developers capture streaming depth data from the camera at a limited resolution. This sample app is the basis for my solution.
Using both an iPhone X running AVCamPhotoFilter and an iPhone 7 Plus in the standard camera application, I stacked the devices, keeping the lenses as close together as possible. I recorded the same scene on each, screen capturing the depth data on my iPhone X to make a moving depth map. The video below demonstrates the process involved and the depth effect that results.
Bringing both pieces of footage into Adobe After Effects, I was able to add shallow depth of field quite easily to my video by applying a camera lens blur to my footage, and telling After Effects the depth map source. This is essentially what iOS does with depth data today, just behind the scenes.
The result isn’t perfect, but took a fraction of the time that building a depth map by hand would take. The output would be more accurate if both the source video and depth data came from the same camera lens, but AVCamPhotoFilter doesn’t support capturing both concurrently. While there are significant limitations to this workaround, the end result is surprisingly almost as polished as the depth effect is on still photos.
Unlocking Portrait Mode for video on current iPhone hardware may prove to be challenging. The feature is already computationally intensive for still photos, and would be significantly more taxing in a video. Third-party applications like Fabby have attempted to recreate the effect entirely in software, but aren’t convincing. However, Apple’s A-series chips and camera hardware continue to advance by leaps and bounds on a yearly basis, so this feature might not be too far out of reach.
The possibilities of Portrait mode videos extend far beyond simple shallow depth of field effects. The same data could eventually be used to simulate Portrait Lighting in videos – just like in Apple’s own TV ad.
Creative manipulation of depth data could even make possible effects like tilt-shift videos and instant masking of subjects as if they were standing in front of a green screen. Apple has already used this technique to great effect in a recent update to their Clips app on iPhone X, adding “Selfie Scenes” that can place you downtown in a city or even on the set of Star Wars.
Finally, Portrait Mode for video could further establish the iPhone as an essential filmmaking tool. Traditional cameras do not capture depth data at all, giving the iPhone an immediate advantage over even high-end video gear.
Apple has made their dedication to the iPhone’s camera clear, funding a short film shot entirely on iPhone, and devoting significant engineering resources to new camera features with every new model. Portrait Mode videos would take the iPhone one step closer to the goal of not only being the best camera you have with you, but the best camera, period.
Joby Gorillapod Zoom Tripod:
Sunpak Handheld Stabilizer:
Cam Caddie Universal Mount For GoPro:
Lume Cube GoPro Mounting Bar:
GoPro 3.5mm Mic Adapter:
Universal Horseshoe Adapter:
MyCharge Adventure Portable Battery Charger:
Rode Video Micro (microphone):
1. GoPro 6 or cell phone (iphone X/Samsung S9+)
2. PULUZ for GoPro HERO6 Toughened Glass Lens + LCD
2. Rode Videomic micro Me / Videomic micro
The Videomic Me can also be used on cellphones. The Videomic micro comes with a shoe stand and cable while the Videomic me doesn’t (sad story) Both come with the deadcat. The Videomic me might plug directly into the Gopro mic adapter. So if you want to use both on Gopro and phone, you either have to buy both or live with just the Videomic Me and use it like that. The Videomic me also has a rear 3.5 connector for a ligtning to mic adapter to use on an iPhone. No ligtning versions only available as yet or direct usb C to plug in straight into a gopro with a mic adapter
3. GoPro Mic adapter
You need this bulky dongle unfortunately to connect an external mic to a GoPro 6. No way around it was yet.
4. A handheld gimbal
Numerous options are available. For mobile these are the winners:
1. Zhiyun-Tech Smooth 3 Professional 3-Axis Handheld Stabilizer for Smartphone
2. Removu S1
5. A stabilizing plate
For the cellphone gimbals you needs a stabilizing plate as the gopro is heavy and it puts to much strain on the motors:
6. An aluminium housing perhaps with a possibility for a hot shoe adapter:
Or a GoPro supersuit dpending on your needs and protection options.
7. A dual battery charger maybe
8. Extra batteries
9. High capacity SD cards
10. Rode SC7 patch cable if you buy the Rode Videomicro Me version (as not included, only in Rode Videomicro with the stabilizing hot shoe harness)
11. IR vision
Depending on your needs there are also nightvision lenses available from Ragecams (http://stores.ebay.co.uk/RageCams?_trksid=p2047675.l2563). For best use of these you need an IR illuminator and you can’t use these during the day when modded.
If I was to build a two or more board system, it would be same boards and stack them. Something like this
These are for rendering farms mostly, least amount of space, modular computing.
Besides the CAT S60 phone, there are also add ons for cellphones to capture thermal imaging.
Link to indigogo: