Tuesday, September 19, 2017 - 15:03

SMPTE Executive Director Barbara Lange and Director of Engineering and Standards Howard Lukk visit IBCTV at IBC2017 to discusses the emerging SMPTE ST 2110 suite of standards for Professional Media Over Managed IP Networks.

Learn more about SMPTE ST 2110 here.


Tuesday, September 19, 2017 - 14:52

SMPTE’s Barbara Lange discusses how standards are more critical than ever in this technologically advanced and video-driven environment.

Thursday, December 29, 2016 - 12:00

Our Centennial year has been a special time for SMPTE and no matter how you have contributed, you have helped to make 2016 an extraordinary year for the Society!

This video is a thank you for all you helped us accomplish in 2016:

We know it's long - there were so many highlights in 2016! We hope you'll watch it all the way through!

We wish you health, happiness, and joy in 2017! We look forward to SMPTE's Next Century knowing we have your continued support.
Here's to the Next Century - Happy New Year from the SMPTE Team!

SMPTE - The Next Century

Monday, November 28, 2016 - 12:48

We are asking our Members to consider supporting SMPTE on Tuesday, 29 November 2016, by posting an #UNselfie on your Twitter, Instagram, and/or FaceBook.

Coinciding with the U.S. Thanksgiving Holiday and the kickoff of the holiday shopping season, #GivingTuesday inspires people to take collaborative action to give back in better, smarter ways to the charities and causes they support and help create a better world. #GivingTuesday harnesses the power of the media to create a moment around the holidays dedicated to giving. 

Download the #SMPTE100 #UNSelfie Template!

It's easy to show your support of SMPTE:

  1. Download the template by clicking the button above
  2. Print the PDF
  3. In the empty space, write your reasons for supporting SMPTE (if you need some ideas, we've included some examples on page 2 of the template)
  4. Hold up the sign and take your #UNselfie
  5. Post your picture to Twitter, Instagram, and/or Facebook on Tuesday, 29 November 2016

Be sure to include the hashtags #SMPTE100#UNselfie#GivingTuesday and feel free to include a link to too!

Thank you for your support!

With much gratitude,

The SMPTE Team

Monday, October 31, 2016 - 12:54

Produced in Partnership with NewBay Media

NewBay Media


SMPTE's Centennial Hall Offers View of Past, Present and Future
Centennial Hall highlights the organization's continued focus on innovation and collaboration. 
SMPTE: Conjuring the Virtual Media Facility
To cloud or not to cloud… live 
SMPTE: How Can You Ensure an Effective VR/AR Experience?
VR/AR sessions focused primarily on strategies to ensure end users a high-quality consumer experience 

Full Newsletter


SMPTE Keynote Explores the Past and Promising Future of Film and Television
Trumbull and Seidel's opening keynote offered perspective, a look at processes and archival video of many pivotal moments in imaging innovation. 
IP Standards, APIs and Moore’s Law 
Media transport infrastructure standards are mashing up the best of the old and the new
Color Management: Is What You See What You Get?
Color management sessions examined some of the challenges content creators will face, both from technical and creative perspectives

Full Newsletter


The Digital Archive of the Future: What Will Be Involved?
Monday's opening sessions began with a challenging keynote from Daniel Teruggi.
SMPTE: Victoria Alonso —Stopping Traffic, Stepping Up, Speaking Out
Alonso spoke during the Women in Technology Luncheon about the challenges of being a female working in Hollywood.
SMPTE Bestowed Honors and Awards
With a Roaring ’20s theme, a speakeasy and a cabaret, SMPTE presented the 2016 Honors & Awards after its red carpet and dinner reception last night. 

Full Newsletter

Monday, October 31, 2016 - 10:59

Produced in Partnership With Broadcast Beat:

Broadcast Beat Logo

Gala  Honors & Awards  Exhibit Halls  Studio  Women in Tech  Membership Meeting  Home Theater Geeks Wrap-Up

Centennial Gala Playlist:

Hear from James Cameron, Billy Zane, John O'Hurley, Richard Edlund, Miles Brown, Marsai Martin, and Barbara Lange live from the SMPTE Centennial Gala arrival red carpet!

SMPTE 2016 Honors & Awards Video Playlist:

Learn more about the SMPTE 2016 Award Recipients and Leadership! More videos being added!

Broadcast Beat Live from the Exhibit Floors Playlist:

Check back for more videos!

SMPTE 2016 Live! Studio Playlist:

More videos to come - please check back!

Annual General Membership Meeting (full video):

Women in Technology Luncheon



Home Theater Geeks SMPTE 2016 Episode:


Tuesday, June 28, 2016 - 18:06

Annie Chang of Walt Disney moderated a broad panel on this hot button topic. For many attendees HDR and wide-color gamut are more important than 4K resolution to image quality -- but they have received much less attention in the media and from consumers. Matthew Goldman, from Ericsson, started us off by defining HDR -- first as one of the five key technologies of UltraHD. However, he said many people use HDR (or HDR+) to mean High Dynamic Range plus Wide Color Gamut (WGC), and 10-bit sampling. Whichever of these definitions is used, he wanted to make it clear that it was not resolution dependent, and specifically does not require 4K resolution. That's particularly important in mobile where small screens might still be able to take advantage of HDR. While it might go without saying, Goldman also reiterated that HDR doesn't mean a brighter display -- it means more contrast. Using charts of color gamuts for both HD (Rec 709) & UltraHD (Rec. 2020) Goldman demonstrated that the combination of HDR & WGC were closely-linked in creating a better image.

Michael Zink, from Warner Bros., walked us through an analysis of the movie LEGO in HDR, and showed that high light levels were only used ina small number of frames, and only for certain highlights. So the HDR version of the movie wasn't overall significantly brighter than the SDR version.

Steven Robertson, Google software engineer, pointed out that content streamed to computers and devices have some major advantages in adapting to new technologies. As devices happen to come with higher-resolution or wider-gamut displays the content can be automatically adapted as it is delivered. This is particularly relevant since many of the latest phones have very similar quality displays to the high-end TVs. Robin Atkins, from Dolby Labs, agreed with Roberson, and said they have found that Vudu and other OTT services were able to incorporate new technologies very rapidly since they had end to end control over their content delivery.

Panelists agreed that another challenge was viewing conditions, both optimizing them, and using metadata to adapt the content stream to adapt to them where it is possible to determine what they are. Atkins also walked us through how dynamic (content-dependent) metadata can be used to assist in mapping the color volume (gamut and tone) of the input scene to particular displays. For example, some scenes might be designed to be dark (perhaps a night-time scene), and shouldn't be mapped into a more typical rendering, while others (perhaps a discussion between characters where facial expressions are important) might be designed to always be easily viewable and should be remapped. There are already over 50 titles available for streaming that use Dolby Vision's "Smart Content" dynamic metadata.

Tuesday, June 28, 2016 - 16:32

The conference's Day 2 keynote was given by Ron Sanders, CEO of Warner Home Entertainment, in the form of an interview conducted by  EnGadget's Roberto Baldwin. Ron started off by confirming what others had said -- that change is accelerating in the entertainment business. As fast as technology improves, customer expectations are increasing just as rapidly. The proliferation of devices has made the job of meeting those expectations much more difficult, but worth it as the overall demand for content continues to increase. 

On the topic of the profusion of new video technologies like 4K, UltraHD, HDR, wider color gamut, and Dolby Vision, Sanders said the first problem is deciding which of those formats are even worth promoting, but the improved viewing experience of all these technologies together is certainly justified by an improved viewing experience. Consistent messaging of the benefits, as well as refitting the retail purchase experience to showcase the new features. Warner is promoting versions of its movies -- about 100 this year -- that include a range of these features, along with immersive audio, as "4K UltraHD." To make these versions practical, they need to be considered early in the production process -- and part of what the Director thinks about -- especially since the theater-only window continues to shrink, meaning there is a need for shorter time-to-market for home releases. Sanders compared the difference between 1080p TV and 4K UltraHD being as large as the upgrade from VHS to DVD. Short term Sanders sees the new versions as a complement to the existing HD & SD versions, but over time more of the market will move to the higher-quality versions. Industry group DEG (Digital Equipment Group), chaired by Sanders, is trying to coordinate consistent messaging around the rollout of the new technologies.

Tuesday, June 28, 2016 - 16:01

Switching from video to audio, Sunil Bharitkar from HP led a panel on the challenges and future direction of audio as part of entertainment. Underlying the changes is a massive upgrade in the delivery technologies for audio, from VHS to DVD and now potentially very-high-bandwidth internet connections. The devices used to play the content has also improved in general, from mono to stereo, to mutlichannel. Mobile devices were a step backwards in many cases, but even those now have access to fairly-high quality audio reproduction. VR represents the next frontier, as spatial audio is a key component of realistic immersive experiences. 

Robert Fisher, Warner Brothers, traced the history of high-quality sound production from studio to cinema, and explained how they have been further extending their production pipeline to accommodate streaming and mobile devices. He stressed that WB wants to ensure that the tradition of excellent sound capture and production is maintained going forward, and called on the industry to help make that happen. As part of this effort, WB has been working with Dolby and others to create the right tools to create a high-quality remix for binaural headphones. Fisher reported that the early results of the new binaural re-mixing technology have been very impressive.

Tim Carroll, Dolby Labs, said there has been great progress in codec quality over the last year, but end-to-end system reproduction remains a problem, as distribution pipelines and players have such a wide range. Just because something sounds good in multi-channel theater audio doesn't automatically mean that it is the right mix for a pair of mobile headphones. This tied-in well with Fisher's call for doing binaural-specific remixes of content.

Deep Sen, Qualcomm, took the audience through how scene-based audio works, and why it is particularly important for VR. In particular, VR drives both a requirement for a true 3D sound field, and to be able to rotate that sound field in response to head movements. Object-based capture is very difficult in these environments, as there can be many sound sources, and they can also be moving. Scene-based Audio is a little like light field capture for video -- it includes sound pressure measurements based on space and time. A specialized array of microphones that captures sound from every direction is used, and the result is mixed with any desired object-based audio sources. Qualcomm helps support the production of scene-based audio by providing plugins for ProTools. The output can be rendered in MPEG-H format for use with VR. To support head tracking the final result also has to be dynamically rotated in real time -- with less than 50ms latency -- to support motion and head tracking. That is in addition to the initial rendering through a standard or personalized HRTF (Head Response Transfer Function). 

Phil Hilmes, Amazon Lab 126, addressed the progress and challenges of playing content in a consumer environment. Those range from synchronization across multiple devices and rooms -- for both audio and video, as well as with sensor and camera data. While there are vendor-specific solutions like Sonos's sound synchronization, Hilmes called for an industry standards effort so that a variety of devices from different vendors can all be synchronized with the user's needs. Sensor input can also be used to help optimize and personalize the audio experience -- especially if it can be standardized and shared. 

The panel discussed that arrays of simple devices, like the Echo, that could be calibrated, synced, and used to render a sound field, might be an alternative to more-traditional fixed -- and often hard to setup correctly -- multi-channel speaker arrays. Multiple sound bars used in combination was also discussed as an option.


Tuesday, June 28, 2016 - 13:21

Just as an increasing amount of content is being consumed on mobile devices, so too is the car a growth market for entertainment -- although one that has its own unique technical and safety issues. Allan McLennan of The PADEM Group moderated a panel including GM's Allan Messer, Michelle Avary from Aeris, and Marc Manus, a consultant to the entertainment industry. Messer kicked things off by stressing the upcoming explosive changes in the automotive industry -- comparing the changes expected over the next five years to the changes from the last 50. His vision video was focused on GM's well-known On Star connectivity platform, and its various applications in security, vehicle management, safety, mobile data access, and entertainment. He positioned connectivity as one of the four major themes driving changes to autos -- Electric propulsion, Sharing, and Autonomous operation being the other three.

Avary, formerly from Toyota, gave some background on Aeris, that brings connectivity and services to devices -- including cars. Their focus is primarily on vehicle information, more so than entertainment. She said that about half the cars being built in the US this year will have an embedded phone (distinct from those that can be used with a tethered smartphone). How many of those customers will actually subscribe to the optional services available for the car's embedded phone after a trial period was a subject for speculation -- with none of the panelists hazarding a specific estimate. Manus said that so far he'd seen the most interest from Hollywood in using the connected experience to help market and generate brand awareness -- through long-form commercials that merge a scripted narrative with a product. The panel tried to hash out how large the market needed to be before content is created specifically for consumption in the car. Right now, driver content is limited to audio, so radio and related models are leading the way. Most video in the car is simply the same content that might be consumed elsewhere, but either downloaded or streamed into the vehicle. McClennan said some of the MCNs in the UK are specifically targeting kid's and the car "back seat" for their programming. Panelists conjectured that when content is created for viewing in cars -- particularly shared cars -- it would need to be authored and produced differently from versions made for consumption in the home. As an example, Avary asked what might be possible if content creators could have access to the car's sensors, and incorporate their input into the experience (anything from location-based informational content to driving-related experiences based on vehicle speed and dynamics).