diff options
Diffstat (limited to 'isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/.svn/text-base/FlatManifestWriterImpl.java.svn-base')
-rw-r--r-- | isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/.svn/text-base/FlatManifestWriterImpl.java.svn-base | 643 |
1 files changed, 0 insertions, 643 deletions
diff --git a/isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/.svn/text-base/FlatManifestWriterImpl.java.svn-base b/isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/.svn/text-base/FlatManifestWriterImpl.java.svn-base deleted file mode 100644 index 5cc9be9..0000000 --- a/isoparser/src/main/java/com/googlecode/mp4parser/authoring/adaptivestreaming/.svn/text-base/FlatManifestWriterImpl.java.svn-base +++ /dev/null @@ -1,643 +0,0 @@ -/* - * Copyright 2012 Sebastian Annies, Hamburg - * - * Licensed under the Apache License, Version 2.0 (the License); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an AS IS BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package com.googlecode.mp4parser.authoring.adaptivestreaming; - -import com.coremedia.iso.Hex; -import com.coremedia.iso.boxes.SampleDescriptionBox; -import com.coremedia.iso.boxes.SoundMediaHeaderBox; -import com.coremedia.iso.boxes.VideoMediaHeaderBox; -import com.coremedia.iso.boxes.h264.AvcConfigurationBox; -import com.coremedia.iso.boxes.sampleentry.AudioSampleEntry; -import com.coremedia.iso.boxes.sampleentry.VisualSampleEntry; -import com.googlecode.mp4parser.Version; -import com.googlecode.mp4parser.authoring.Movie; -import com.googlecode.mp4parser.authoring.Track; -import com.googlecode.mp4parser.authoring.builder.FragmentIntersectionFinder; -import com.googlecode.mp4parser.boxes.DTSSpecificBox; -import com.googlecode.mp4parser.boxes.EC3SpecificBox; -import com.googlecode.mp4parser.boxes.mp4.ESDescriptorBox; -import com.googlecode.mp4parser.boxes.mp4.objectdescriptors.AudioSpecificConfig; -import org.w3c.dom.Document; -import org.w3c.dom.Element; - -import javax.xml.parsers.DocumentBuilder; -import javax.xml.parsers.DocumentBuilderFactory; -import javax.xml.parsers.ParserConfigurationException; -import javax.xml.transform.*; -import javax.xml.transform.dom.DOMSource; -import javax.xml.transform.stream.StreamResult; -import java.io.ByteArrayOutputStream; -import java.io.IOException; -import java.io.StringWriter; -import java.nio.ByteBuffer; -import java.util.LinkedList; -import java.util.List; -import java.util.logging.Logger; - -public class FlatManifestWriterImpl extends AbstractManifestWriter { - private static final Logger LOG = Logger.getLogger(FlatManifestWriterImpl.class.getName()); - - protected FlatManifestWriterImpl(FragmentIntersectionFinder intersectionFinder) { - super(intersectionFinder); - } - - /** - * Overwrite this method in subclasses to add your specialities. - * - * @param manifest the original manifest - * @return your customized version of the manifest - */ - protected Document customizeManifest(Document manifest) { - return manifest; - } - - public String getManifest(Movie movie) throws IOException { - - LinkedList<VideoQuality> videoQualities = new LinkedList<VideoQuality>(); - long videoTimescale = -1; - - LinkedList<AudioQuality> audioQualities = new LinkedList<AudioQuality>(); - long audioTimescale = -1; - - for (Track track : movie.getTracks()) { - if (track.getMediaHeaderBox() instanceof VideoMediaHeaderBox) { - videoFragmentsDurations = checkFragmentsAlign(videoFragmentsDurations, calculateFragmentDurations(track, movie)); - SampleDescriptionBox stsd = track.getSampleDescriptionBox(); - videoQualities.add(getVideoQuality(track, (VisualSampleEntry) stsd.getSampleEntry())); - if (videoTimescale == -1) { - videoTimescale = track.getTrackMetaData().getTimescale(); - } else { - assert videoTimescale == track.getTrackMetaData().getTimescale(); - } - } - if (track.getMediaHeaderBox() instanceof SoundMediaHeaderBox) { - audioFragmentsDurations = checkFragmentsAlign(audioFragmentsDurations, calculateFragmentDurations(track, movie)); - SampleDescriptionBox stsd = track.getSampleDescriptionBox(); - audioQualities.add(getAudioQuality(track, (AudioSampleEntry) stsd.getSampleEntry())); - if (audioTimescale == -1) { - audioTimescale = track.getTrackMetaData().getTimescale(); - } else { - assert audioTimescale == track.getTrackMetaData().getTimescale(); - } - - } - } - DocumentBuilderFactory documentBuilderFactory = DocumentBuilderFactory.newInstance(); - DocumentBuilder documentBuilder; - try { - documentBuilder = documentBuilderFactory.newDocumentBuilder(); - } catch (ParserConfigurationException e) { - throw new IOException(e); - } - Document document = documentBuilder.newDocument(); - - - Element smoothStreamingMedia = document.createElement("SmoothStreamingMedia"); - document.appendChild(smoothStreamingMedia); - smoothStreamingMedia.setAttribute("MajorVersion", "2"); - smoothStreamingMedia.setAttribute("MinorVersion", "1"); -// silverlight ignores the timescale attr smoothStreamingMedia.addAttribute(new Attribute("TimeScale", Long.toString(movieTimeScale))); - smoothStreamingMedia.setAttribute("Duration", "0"); - - smoothStreamingMedia.appendChild(document.createComment(Version.VERSION)); - Element videoStreamIndex = document.createElement("StreamIndex"); - videoStreamIndex.setAttribute("Type", "video"); - videoStreamIndex.setAttribute("TimeScale", Long.toString(videoTimescale)); // silverlight ignores the timescale attr - videoStreamIndex.setAttribute("Chunks", Integer.toString(videoFragmentsDurations.length)); - videoStreamIndex.setAttribute("Url", "video/{bitrate}/{start time}"); - videoStreamIndex.setAttribute("QualityLevels", Integer.toString(videoQualities.size())); - smoothStreamingMedia.appendChild(videoStreamIndex); - - for (int i = 0; i < videoQualities.size(); i++) { - VideoQuality vq = videoQualities.get(i); - Element qualityLevel = document.createElement("QualityLevel"); - qualityLevel.setAttribute("Index", Integer.toString(i)); - qualityLevel.setAttribute("Bitrate", Long.toString(vq.bitrate)); - qualityLevel.setAttribute("FourCC", vq.fourCC); - qualityLevel.setAttribute("MaxWidth", Long.toString(vq.width)); - qualityLevel.setAttribute("MaxHeight", Long.toString(vq.height)); - qualityLevel.setAttribute("CodecPrivateData", vq.codecPrivateData); - qualityLevel.setAttribute("NALUnitLengthField", Integer.toString(vq.nalLength)); - videoStreamIndex.appendChild(qualityLevel); - } - - for (int i = 0; i < videoFragmentsDurations.length; i++) { - Element c = document.createElement("c"); - c.setAttribute("n", Integer.toString(i)); - c.setAttribute("d", Long.toString(videoFragmentsDurations[i])); - videoStreamIndex.appendChild(c); - } - - if (audioFragmentsDurations != null) { - Element audioStreamIndex = document.createElement("StreamIndex"); - audioStreamIndex.setAttribute("Type", "audio"); - audioStreamIndex.setAttribute("TimeScale", Long.toString(audioTimescale)); // silverlight ignores the timescale attr - audioStreamIndex.setAttribute("Chunks", Integer.toString(audioFragmentsDurations.length)); - audioStreamIndex.setAttribute("Url", "audio/{bitrate}/{start time}"); - audioStreamIndex.setAttribute("QualityLevels", Integer.toString(audioQualities.size())); - smoothStreamingMedia.appendChild(audioStreamIndex); - - for (int i = 0; i < audioQualities.size(); i++) { - AudioQuality aq = audioQualities.get(i); - Element qualityLevel = document.createElement("QualityLevel"); - qualityLevel.setAttribute("Index", Integer.toString(i)); - qualityLevel.setAttribute("FourCC", aq.fourCC); - qualityLevel.setAttribute("Bitrate", Long.toString(aq.bitrate)); - qualityLevel.setAttribute("AudioTag", Integer.toString(aq.audioTag)); - qualityLevel.setAttribute("SamplingRate", Long.toString(aq.samplingRate)); - qualityLevel.setAttribute("Channels", Integer.toString(aq.channels)); - qualityLevel.setAttribute("BitsPerSample", Integer.toString(aq.bitPerSample)); - qualityLevel.setAttribute("PacketSize", Integer.toString(aq.packetSize)); - qualityLevel.setAttribute("CodecPrivateData", aq.codecPrivateData); - audioStreamIndex.appendChild(qualityLevel); - } - for (int i = 0; i < audioFragmentsDurations.length; i++) { - Element c = document.createElement("c"); - c.setAttribute("n", Integer.toString(i)); - c.setAttribute("d", Long.toString(audioFragmentsDurations[i])); - audioStreamIndex.appendChild(c); - } - } - - document.setXmlStandalone(true); - Source source = new DOMSource(document); - StringWriter stringWriter = new StringWriter(); - Result result = new StreamResult(stringWriter); - TransformerFactory factory = TransformerFactory.newInstance(); - Transformer transformer; - try { - transformer = factory.newTransformer(); - transformer.setOutputProperty(OutputKeys.INDENT, "yes"); - transformer.transform(source, result); - } catch (TransformerConfigurationException e) { - throw new IOException(e); - } catch (TransformerException e) { - throw new IOException(e); - } - return stringWriter.getBuffer().toString(); - - - } - - private AudioQuality getAudioQuality(Track track, AudioSampleEntry ase) { - if (getFormat(ase).equals("mp4a")) { - return getAacAudioQuality(track, ase); - } else if (getFormat(ase).equals("ec-3")) { - return getEc3AudioQuality(track, ase); - } else if (getFormat(ase).startsWith("dts")) { - return getDtsAudioQuality(track, ase); - } else { - throw new InternalError("I don't know what to do with audio of type " + getFormat(ase)); - } - - } - - private AudioQuality getAacAudioQuality(Track track, AudioSampleEntry ase) { - AudioQuality l = new AudioQuality(); - final ESDescriptorBox esDescriptorBox = ase.getBoxes(ESDescriptorBox.class).get(0); - final AudioSpecificConfig audioSpecificConfig = esDescriptorBox.getEsDescriptor().getDecoderConfigDescriptor().getAudioSpecificInfo(); - if (audioSpecificConfig.getSbrPresentFlag() == 1) { - l.fourCC = "AACH"; - } else if (audioSpecificConfig.getPsPresentFlag() == 1) { - l.fourCC = "AACP"; //I'm not sure if that's what MS considers as AAC+ - because actually AAC+ and AAC-HE should be the same... - } else { - l.fourCC = "AACL"; - } - l.bitrate = getBitrate(track); - l.audioTag = 255; - l.samplingRate = ase.getSampleRate(); - l.channels = ase.getChannelCount(); - l.bitPerSample = ase.getSampleSize(); - l.packetSize = 4; - l.codecPrivateData = getAudioCodecPrivateData(audioSpecificConfig); - //Index="0" Bitrate="103000" AudioTag="255" SamplingRate="44100" Channels="2" BitsPerSample="16" packetSize="4" CodecPrivateData="" - return l; - } - - private AudioQuality getEc3AudioQuality(Track track, AudioSampleEntry ase) { - final EC3SpecificBox ec3SpecificBox = ase.getBoxes(EC3SpecificBox.class).get(0); - if (ec3SpecificBox == null) { - throw new RuntimeException("EC-3 track misses EC3SpecificBox!"); - } - - short nfchans = 0; //full bandwidth channels - short lfechans = 0; - byte dWChannelMaskFirstByte = 0; - byte dWChannelMaskSecondByte = 0; - for (EC3SpecificBox.Entry entry : ec3SpecificBox.getEntries()) { - /* - Table 4.3: Audio coding mode - acmod Audio coding mode Nfchans Channel array ordering - 000 1 + 1 2 Ch1, Ch2 - 001 1/0 1 C - 010 2/0 2 L, R - 011 3/0 3 L, C, R - 100 2/1 3 L, R, S - 101 3/1 4 L, C, R, S - 110 2/2 4 L, R, SL, SR - 111 3/2 5 L, C, R, SL, SR - - Table F.2: Chan_loc field bit assignments - Bit Location - 0 Lc/Rc pair - 1 Lrs/Rrs pair - 2 Cs - 3 Ts - 4 Lsd/Rsd pair - 5 Lw/Rw pair - 6 Lvh/Rvh pair - 7 Cvh - 8 LFE2 - */ - switch (entry.acmod) { - case 0: //1+1; Ch1, Ch2 - nfchans += 2; - throw new RuntimeException("Smooth Streaming doesn't support DDP 1+1 mode"); - case 1: //1/0; C - nfchans += 1; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0x20; - } - break; - case 2: //2/0; L, R - nfchans += 2; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xC0; - } - break; - case 3: //3/0; L, C, R - nfchans += 3; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xE0; - } - break; - case 4: //2/1; L, R, S - nfchans += 3; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xC0; - dWChannelMaskSecondByte |= 0x80; - } - break; - case 5: //3/1; L, C, R, S - nfchans += 4; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xE0; - dWChannelMaskSecondByte |= 0x80; - } - break; - case 6: //2/2; L, R, SL, SR - nfchans += 4; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xCC; - } - break; - case 7: //3/2; L, C, R, SL, SR - nfchans += 5; - if (entry.num_dep_sub > 0) { - DependentSubstreamMask dependentSubstreamMask = new DependentSubstreamMask(dWChannelMaskFirstByte, dWChannelMaskSecondByte, entry).process(); - dWChannelMaskFirstByte |= dependentSubstreamMask.getdWChannelMaskFirstByte(); - dWChannelMaskSecondByte |= dependentSubstreamMask.getdWChannelMaskSecondByte(); - } else { - dWChannelMaskFirstByte |= 0xEC; - } - break; - } - if (entry.lfeon == 1) { - lfechans ++; - dWChannelMaskFirstByte |= 0x10; - } - } - - final ByteBuffer waveformatex = ByteBuffer.allocate(22); - waveformatex.put(new byte[]{0x00, 0x06}); //1536 wSamplesPerBlock - little endian - waveformatex.put(dWChannelMaskFirstByte); - waveformatex.put(dWChannelMaskSecondByte); - waveformatex.put(new byte[]{0x00, 0x00}); //pad dwChannelMask to 32bit - waveformatex.put(new byte[]{(byte)0xAF, (byte)0x87, (byte)0xFB, (byte)0xA7, 0x02, 0x2D, (byte)0xFB, 0x42, (byte)0xA4, (byte)0xD4, 0x05, (byte)0xCD, (byte)0x93, (byte)0x84, 0x3B, (byte)0xDD}); //SubFormat - Dolby Digital Plus GUID - - final ByteBuffer dec3Content = ByteBuffer.allocate((int) ec3SpecificBox.getContentSize()); - ec3SpecificBox.getContent(dec3Content); - - AudioQuality l = new AudioQuality(); - l.fourCC = "EC-3"; - l.bitrate = getBitrate(track); - l.audioTag = 65534; - l.samplingRate = ase.getSampleRate(); - l.channels = nfchans + lfechans; - l.bitPerSample = 16; - l.packetSize = track.getSamples().get(0).limit(); //assuming all are same size - l.codecPrivateData = Hex.encodeHex(waveformatex.array()) + Hex.encodeHex(dec3Content.array()); //append EC3SpecificBox (big endian) at the end of waveformatex - return l; - } - - private AudioQuality getDtsAudioQuality(Track track, AudioSampleEntry ase) { - final DTSSpecificBox dtsSpecificBox = ase.getBoxes(DTSSpecificBox.class).get(0); - if (dtsSpecificBox == null) { - throw new RuntimeException("DTS track misses DTSSpecificBox!"); - } - - final ByteBuffer waveformatex = ByteBuffer.allocate(22); - final int frameDuration = dtsSpecificBox.getFrameDuration(); - short samplesPerBlock = 0; - switch (frameDuration) { - case 0: - samplesPerBlock = 512; - break; - case 1: - samplesPerBlock = 1024; - break; - case 2: - samplesPerBlock = 2048; - break; - case 3: - samplesPerBlock = 4096; - break; - } - waveformatex.put((byte) (samplesPerBlock & 0xff)); - waveformatex.put((byte) (samplesPerBlock >>> 8)); - final int dwChannelMask = getNumChannelsAndMask(dtsSpecificBox)[1]; - waveformatex.put((byte) (dwChannelMask & 0xff)); - waveformatex.put((byte) (dwChannelMask >>> 8)); - waveformatex.put((byte) (dwChannelMask >>> 16)); - waveformatex.put((byte) (dwChannelMask >>> 24)); - waveformatex.put(new byte[]{(byte)0xAE, (byte)0xE4, (byte)0xBF, (byte)0x5E, (byte)0x61, (byte)0x5E, (byte)0x41, (byte)0x87, (byte)0x92, (byte)0xFC, (byte)0xA4, (byte)0x81, (byte)0x26, (byte)0x99, (byte)0x02, (byte)0x11}); //DTS-HD GUID - - final ByteBuffer dtsCodecPrivateData = ByteBuffer.allocate(8); - dtsCodecPrivateData.put((byte) dtsSpecificBox.getStreamConstruction()); - - final int channelLayout = dtsSpecificBox.getChannelLayout(); - dtsCodecPrivateData.put((byte) (channelLayout & 0xff)); - dtsCodecPrivateData.put((byte) (channelLayout >>> 8)); - dtsCodecPrivateData.put((byte) (channelLayout >>> 16)); - dtsCodecPrivateData.put((byte) (channelLayout >>> 24)); - - byte dtsFlags = (byte) (dtsSpecificBox.getMultiAssetFlag() << 1); - dtsFlags |= dtsSpecificBox.getLBRDurationMod(); - dtsCodecPrivateData.put(dtsFlags); - dtsCodecPrivateData.put(new byte[]{0x00, 0x00}); //reserved - - AudioQuality l = new AudioQuality(); - l.fourCC = getFormat(ase); - l.bitrate = dtsSpecificBox.getAvgBitRate(); - l.audioTag = 65534; - l.samplingRate = dtsSpecificBox.getDTSSamplingFrequency(); - l.channels = getNumChannelsAndMask(dtsSpecificBox)[0]; - l.bitPerSample = 16; - l.packetSize = track.getSamples().get(0).limit(); //assuming all are same size - l.codecPrivateData = Hex.encodeHex(waveformatex.array()) + Hex.encodeHex(dtsCodecPrivateData.array()); - return l; - - } - - /* dwChannelMask - L SPEAKER_FRONT_LEFT 0x00000001 - R SPEAKER_FRONT_RIGHT 0x00000002 - C SPEAKER_FRONT_CENTER 0x00000004 - LFE1 SPEAKER_LOW_FREQUENCY 0x00000008 - Ls or Lsr* SPEAKER_BACK_LEFT 0x00000010 - Rs or Rsr* SPEAKER_BACK_RIGHT 0x00000020 - Lc SPEAKER_FRONT_LEFT_OF_CENTER 0x00000040 - Rc SPEAKER_FRONT_RIGHT_OF_CENTER 0x00000080 - Cs SPEAKER_BACK_CENTER 0x00000100 - Lss SPEAKER_SIDE_LEFT 0x00000200 - Rss SPEAKER_SIDE_RIGHT 0x00000400 - Oh SPEAKER_TOP_CENTER 0x00000800 - Lh SPEAKER_TOP_FRONT_LEFT 0x00001000 - Ch SPEAKER_TOP_FRONT_CENTER 0x00002000 - Rh SPEAKER_TOP_FRONT_RIGHT 0x00004000 - Lhr SPEAKER_TOP_BACK_LEFT 0x00008000 - Chf SPEAKER_TOP_BACK_CENTER 0x00010000 - Rhr SPEAKER_TOP_BACK_RIGHT 0x00020000 - SPEAKER_RESERVED 0x80000000 - - * if Lss, Rss exist, then this position is equivalent to Lsr, Rsr respectively - */ - private int[] getNumChannelsAndMask(DTSSpecificBox dtsSpecificBox) { - final int channelLayout = dtsSpecificBox.getChannelLayout(); - int numChannels = 0; - int dwChannelMask = 0; - if ((channelLayout & 0x0001) == 0x0001) { - //0001h Center in front of listener 1 - numChannels += 1; - dwChannelMask |= 0x00000004; //SPEAKER_FRONT_CENTER - } - if ((channelLayout & 0x0002) == 0x0002) { - //0002h Left/Right in front 2 - numChannels += 2; - dwChannelMask |= 0x00000001; //SPEAKER_FRONT_LEFT - dwChannelMask |= 0x00000002; //SPEAKER_FRONT_RIGHT - } - if ((channelLayout & 0x0004) == 0x0004) { - //0004h Left/Right surround on side in rear 2 - numChannels += 2; - //* if Lss, Rss exist, then this position is equivalent to Lsr, Rsr respectively - dwChannelMask |= 0x00000010; //SPEAKER_BACK_LEFT - dwChannelMask |= 0x00000020; //SPEAKER_BACK_RIGHT - } - if ((channelLayout & 0x0008) == 0x0008) { - //0008h Low frequency effects subwoofer 1 - numChannels += 1; - dwChannelMask |= 0x00000008; //SPEAKER_LOW_FREQUENCY - } - if ((channelLayout & 0x0010) == 0x0010) { - //0010h Center surround in rear 1 - numChannels += 1; - dwChannelMask |= 0x00000100; //SPEAKER_BACK_CENTER - } - if ((channelLayout & 0x0020) == 0x0020) { - //0020h Left/Right height in front 2 - numChannels += 2; - dwChannelMask |= 0x00001000; //SPEAKER_TOP_FRONT_LEFT - dwChannelMask |= 0x00004000; //SPEAKER_TOP_FRONT_RIGHT - } - if ((channelLayout & 0x0040) == 0x0040) { - //0040h Left/Right surround in rear 2 - numChannels += 2; - dwChannelMask |= 0x00000010; //SPEAKER_BACK_LEFT - dwChannelMask |= 0x00000020; //SPEAKER_BACK_RIGHT - } - if ((channelLayout & 0x0080) == 0x0080) { - //0080h Center Height in front 1 - numChannels += 1; - dwChannelMask |= 0x00002000; //SPEAKER_TOP_FRONT_CENTER - } - if ((channelLayout & 0x0100) == 0x0100) { - //0100h Over the listener’s head 1 - numChannels += 1; - dwChannelMask |= 0x00000800; //SPEAKER_TOP_CENTER - } - if ((channelLayout & 0x0200) == 0x0200) { - //0200h Between left/right and center in front 2 - numChannels += 2; - dwChannelMask |= 0x00000040; //SPEAKER_FRONT_LEFT_OF_CENTER - dwChannelMask |= 0x00000080; //SPEAKER_FRONT_RIGHT_OF_CENTER - } - if ((channelLayout & 0x0400) == 0x0400) { - //0400h Left/Right on side in front 2 - numChannels += 2; - dwChannelMask |= 0x00000200; //SPEAKER_SIDE_LEFT - dwChannelMask |= 0x00000400; //SPEAKER_SIDE_RIGHT - } - if ((channelLayout & 0x0800) == 0x0800) { - //0800h Left/Right surround on side 2 - numChannels += 2; - //* if Lss, Rss exist, then this position is equivalent to Lsr, Rsr respectively - dwChannelMask |= 0x00000010; //SPEAKER_BACK_LEFT - dwChannelMask |= 0x00000020; //SPEAKER_BACK_RIGHT - } - if ((channelLayout & 0x1000) == 0x1000) { - //1000h Second low frequency effects subwoofer 1 - numChannels += 1; - dwChannelMask |= 0x00000008; //SPEAKER_LOW_FREQUENCY - } - if ((channelLayout & 0x2000) == 0x2000) { - //2000h Left/Right height on side 2 - numChannels += 2; - dwChannelMask |= 0x00000010; //SPEAKER_BACK_LEFT - dwChannelMask |= 0x00000020; //SPEAKER_BACK_RIGHT - } - if ((channelLayout & 0x4000) == 0x4000) { - //4000h Center height in rear 1 - numChannels += 1; - dwChannelMask |= 0x00010000; //SPEAKER_TOP_BACK_CENTER - } - if ((channelLayout & 0x8000) == 0x8000) { - //8000h Left/Right height in rear 2 - numChannels += 2; - dwChannelMask |= 0x00008000; //SPEAKER_TOP_BACK_LEFT - dwChannelMask |= 0x00020000; //SPEAKER_TOP_BACK_RIGHT - } - if ((channelLayout & 0x10000) == 0x10000) { - //10000h Center below in front - numChannels += 1; - } - if ((channelLayout & 0x20000) == 0x20000) { - //20000h Left/Right below in front - numChannels += 2; - } - return new int[]{numChannels, dwChannelMask}; - } - - private String getAudioCodecPrivateData(AudioSpecificConfig audioSpecificConfig) { - byte[] configByteArray = audioSpecificConfig.getConfigBytes(); - return Hex.encodeHex(configByteArray); - } - - private VideoQuality getVideoQuality(Track track, VisualSampleEntry vse) { - VideoQuality l; - if ("avc1".equals(getFormat(vse))) { - AvcConfigurationBox avcConfigurationBox = vse.getBoxes(AvcConfigurationBox.class).get(0); - l = new VideoQuality(); - l.bitrate = getBitrate(track); - l.codecPrivateData = Hex.encodeHex(getAvcCodecPrivateData(avcConfigurationBox)); - l.fourCC = "AVC1"; - l.width = vse.getWidth(); - l.height = vse.getHeight(); - l.nalLength = avcConfigurationBox.getLengthSizeMinusOne() + 1; - } else { - throw new InternalError("I don't know how to handle video of type " + getFormat(vse)); - } - return l; - } - - private byte[] getAvcCodecPrivateData(AvcConfigurationBox avcConfigurationBox) { - List<byte[]> sps = avcConfigurationBox.getSequenceParameterSets(); - List<byte[]> pps = avcConfigurationBox.getPictureParameterSets(); - ByteArrayOutputStream baos = new ByteArrayOutputStream(); - try { - baos.write(new byte[]{0, 0, 0, 1}); - - for (byte[] sp : sps) { - baos.write(sp); - } - baos.write(new byte[]{0, 0, 0, 1}); - for (byte[] pp : pps) { - baos.write(pp); - } - } catch (IOException ex) { - throw new RuntimeException("ByteArrayOutputStream do not throw IOException ?!?!?"); - } - return baos.toByteArray(); - } - - private class DependentSubstreamMask { - private byte dWChannelMaskFirstByte; - private byte dWChannelMaskSecondByte; - private EC3SpecificBox.Entry entry; - - public DependentSubstreamMask(byte dWChannelMaskFirstByte, byte dWChannelMaskSecondByte, EC3SpecificBox.Entry entry) { - this.dWChannelMaskFirstByte = dWChannelMaskFirstByte; - this.dWChannelMaskSecondByte = dWChannelMaskSecondByte; - this.entry = entry; - } - - public byte getdWChannelMaskFirstByte() { - return dWChannelMaskFirstByte; - } - - public byte getdWChannelMaskSecondByte() { - return dWChannelMaskSecondByte; - } - - public DependentSubstreamMask process() { - switch (entry.chan_loc) { - case 0: - dWChannelMaskFirstByte |= 0x3; - break; - case 1: - dWChannelMaskFirstByte |= 0xC; - break; - case 2: - dWChannelMaskSecondByte |= 0x80; - break; - case 3: - dWChannelMaskSecondByte |= 0x8; - break; - case 6: - dWChannelMaskSecondByte |= 0x5; - break; - case 7: - dWChannelMaskSecondByte |= 0x2; - break; - } - return this; - } - } -} |