Unpacking aecdumps generates wav files

BUG=webrtc:3359
R=andrew@webrtc.org, bjornv@webrtc.org, kwiberg@webrtc.org

Review URL: https://webrtc-codereview.appspot.com/14239004

git-svn-id: http://webrtc.googlecode.com/svn/trunk/webrtc@7018 4adac7df-926f-26a2-2b94-8c16560cd09d
diff --git a/common_audio/wav_writer.h b/common_audio/wav_writer.h
index 45bcbac..0966727 100644
--- a/common_audio/wav_writer.h
+++ b/common_audio/wav_writer.h
@@ -33,13 +33,13 @@
   // [-32768,32767], and there must be the previously specified number of
   // interleaved channels.
   void WriteSamples(const float* samples, size_t num_samples);
+  void WriteSamples(const int16_t* samples, size_t num_samples);
 
   int sample_rate() const { return sample_rate_; }
   int num_channels() const { return num_channels_; }
   uint32_t num_samples() const { return num_samples_; }
 
  private:
-  void WriteSamples(const int16_t* samples, size_t num_samples);
   void Close();
   const int sample_rate_;
   const int num_channels_;
diff --git a/modules/audio_processing/audio_processing_tests.gypi b/modules/audio_processing/audio_processing_tests.gypi
index 82aa7fd..99b80f2 100644
--- a/modules/audio_processing/audio_processing_tests.gypi
+++ b/modules/audio_processing/audio_processing_tests.gypi
@@ -41,6 +41,7 @@
           'dependencies': [
             'audioproc_debug_proto',
             '<(webrtc_root)/system_wrappers/source/system_wrappers.gyp:system_wrappers',
+            '<(webrtc_root)/common_audio/common_audio.gyp:common_audio',
             '<(DEPTH)/third_party/gflags/gflags.gyp:gflags',
           ],
           'sources': [ 'test/unpack.cc', ],
diff --git a/modules/audio_processing/test/unpack.cc b/modules/audio_processing/test/unpack.cc
index c90ba82..a225d58 100644
--- a/modules/audio_processing/test/unpack.cc
+++ b/modules/audio_processing/test/unpack.cc
@@ -14,25 +14,28 @@
 // to unpack the file into its component parts: audio and other data.
 
 #include <stdio.h>
+#include <limits>
 
 #include "gflags/gflags.h"
 #include "webrtc/audio_processing/debug.pb.h"
+#include "webrtc/common_audio/include/audio_util.h"
+#include "webrtc/common_audio/wav_writer.h"
 #include "webrtc/modules/audio_processing/test/test_utils.h"
 #include "webrtc/system_wrappers/interface/scoped_ptr.h"
 #include "webrtc/typedefs.h"
 
 // TODO(andrew): unpack more of the data.
 DEFINE_string(input_file, "input.pcm", "The name of the input stream file.");
-DEFINE_string(float_input_file, "input.float",
-              "The name of the float input stream file.");
+DEFINE_string(input_wav_file, "input.wav",
+              "The name of the WAV input stream file.");
 DEFINE_string(output_file, "ref_out.pcm",
               "The name of the reference output stream file.");
-DEFINE_string(float_output_file, "ref_out.float",
-              "The name of the float reference output stream file.");
+DEFINE_string(output_wav_file, "ref_out.wav",
+              "The name of the WAV reference output stream file.");
 DEFINE_string(reverse_file, "reverse.pcm",
               "The name of the reverse input stream file.");
-DEFINE_string(float_reverse_file, "reverse.float",
-              "The name of the float reverse input stream file.");
+DEFINE_string(reverse_wav_file, "reverse.wav",
+              "The name of the WAV reverse input stream file.");
 DEFINE_string(delay_file, "delay.int32", "The name of the delay file.");
 DEFINE_string(drift_file, "drift.int32", "The name of the drift file.");
 DEFINE_string(level_file, "level.int32", "The name of the level file.");
@@ -40,6 +43,7 @@
 DEFINE_string(settings_file, "settings.txt", "The name of the settings file.");
 DEFINE_bool(full, false,
             "Unpack the full set of files (normally not needed).");
+DEFINE_bool(pcm, false, "Write to PCM instead of WAV file.");
 
 namespace webrtc {
 
@@ -48,6 +52,36 @@
 using audioproc::Stream;
 using audioproc::Init;
 
+class PcmFile {
+ public:
+  PcmFile(const std::string& filename)
+      : file_handle_(fopen(filename.c_str(), "wb")) {}
+
+  ~PcmFile() {
+    fclose(file_handle_);
+  }
+
+  void WriteSamples(const int16_t* samples, size_t num_samples) {
+#ifndef WEBRTC_ARCH_LITTLE_ENDIAN
+#error "Need to convert samples to little-endian when writing to PCM file"
+#endif
+    fwrite(samples, sizeof(*samples), num_samples, file_handle_);
+  }
+
+  void WriteSamples(const float* samples, size_t num_samples) {
+    static const size_t kChunksize = 4096 / sizeof(uint16_t);
+    for (size_t i = 0; i < num_samples; i += kChunksize) {
+      int16_t isamples[kChunksize];
+      const size_t chunk = std::min(kChunksize, num_samples - i);
+      RoundToInt16(samples + i, chunk, isamples);
+      WriteSamples(isamples, chunk);
+    }
+  }
+
+ private:
+  FILE* file_handle_;
+};
+
 void WriteData(const void* data, size_t size, FILE* file,
                const std::string& filename) {
   if (fwrite(data, size, 1, file) != 1) {
@@ -56,6 +90,40 @@
   }
 }
 
+void WriteIntData(const int16_t* data,
+                  size_t length,
+                  WavFile* wav_file,
+                  PcmFile* pcm_file) {
+  if (wav_file) {
+    wav_file->WriteSamples(data, length);
+  }
+  if (pcm_file) {
+    pcm_file->WriteSamples(data, length);
+  }
+}
+
+void WriteFloatData(const float* const* data,
+                    size_t samples_per_channel,
+                    int num_channels,
+                    WavFile* wav_file,
+                    PcmFile* pcm_file) {
+  size_t length = num_channels * samples_per_channel;
+  scoped_ptr<float[]> buffer(new float[length]);
+  Interleave(data, samples_per_channel, num_channels, buffer.get());
+  // TODO(aluebs): Use ScaleToInt16Range() from audio_util
+  for (size_t i = 0; i < length; ++i) {
+    buffer[i] = buffer[i] > 0 ?
+                buffer[i] * std::numeric_limits<int16_t>::max() :
+                -buffer[i] * std::numeric_limits<int16_t>::min();
+  }
+  if (wav_file) {
+    wav_file->WriteSamples(buffer.get(), length);
+  }
+  if (pcm_file) {
+    pcm_file->WriteSamples(buffer.get(), length);
+  }
+}
+
 int do_main(int argc, char* argv[]) {
   std::string program_name = argv[0];
   std::string usage = "Commandline tool to unpack audioproc debug files.\n"
@@ -72,7 +140,16 @@
 
   Event event_msg;
   int frame_count = 0;
-while (ReadMessageFromFile(debug_file, &event_msg)) {
+  int num_input_channels = 0;
+  int num_output_channels = 0;
+  int num_reverse_channels = 0;
+  scoped_ptr<WavFile> reverse_wav_file;
+  scoped_ptr<WavFile> input_wav_file;
+  scoped_ptr<WavFile> output_wav_file;
+  scoped_ptr<PcmFile> reverse_pcm_file;
+  scoped_ptr<PcmFile> input_pcm_file;
+  scoped_ptr<PcmFile> output_pcm_file;
+  while (ReadMessageFromFile(debug_file, &event_msg)) {
     if (event_msg.type() == Event::REVERSE_STREAM) {
       if (!event_msg.has_reverse_stream()) {
         printf("Corrupt input file: ReverseStream missing.\n");
@@ -81,17 +158,20 @@
 
       const ReverseStream msg = event_msg.reverse_stream();
       if (msg.has_data()) {
-        static FILE* reverse_file = OpenFile(FLAGS_reverse_file, "wb");
-        WriteData(msg.data().data(), msg.data().size(), reverse_file,
-                  FLAGS_reverse_file);
-
+        WriteIntData(reinterpret_cast<const int16_t*>(msg.data().data()),
+                     msg.data().size() / sizeof(int16_t),
+                     reverse_wav_file.get(),
+                     reverse_pcm_file.get());
       } else if (msg.channel_size() > 0) {
-        static FILE* float_reverse_file = OpenFile(FLAGS_float_reverse_file,
-                                                   "wb");
-        // TODO(ajm): Interleave multiple channels.
-        assert(msg.channel_size() == 1);
-        WriteData(msg.channel(0).data(), msg.channel(0).size(),
-                  float_reverse_file, FLAGS_reverse_file);
+        scoped_ptr<const float*[]> data(new const float*[num_reverse_channels]);
+        for (int i = 0; i < num_reverse_channels; ++i) {
+          data[i] = reinterpret_cast<const float*>(msg.channel(i).data());
+        }
+        WriteFloatData(data.get(),
+                       msg.channel(0).size() / sizeof(float),
+                       num_reverse_channels,
+                       reverse_wav_file.get(),
+                       reverse_pcm_file.get());
       }
     } else if (event_msg.type() == Event::STREAM) {
       frame_count++;
@@ -102,30 +182,38 @@
 
       const Stream msg = event_msg.stream();
       if (msg.has_input_data()) {
-        static FILE* input_file = OpenFile(FLAGS_input_file, "wb");
-        WriteData(msg.input_data().data(), msg.input_data().size(),
-                  input_file, FLAGS_input_file);
-
+        WriteIntData(reinterpret_cast<const int16_t*>(msg.input_data().data()),
+                     msg.input_data().size() / sizeof(int16_t),
+                     input_wav_file.get(),
+                     input_pcm_file.get());
       } else if (msg.input_channel_size() > 0) {
-        static FILE* float_input_file = OpenFile(FLAGS_float_input_file, "wb");
-        // TODO(ajm): Interleave multiple channels.
-        assert(msg.input_channel_size() == 1);
-        WriteData(msg.input_channel(0).data(), msg.input_channel(0).size(),
-                  float_input_file, FLAGS_float_input_file);
+        scoped_ptr<const float*[]> data(new const float*[num_input_channels]);
+        for (int i = 0; i < num_input_channels; ++i) {
+          data[i] = reinterpret_cast<const float*>(msg.input_channel(i).data());
+        }
+        WriteFloatData(data.get(),
+                       msg.input_channel(0).size() / sizeof(float),
+                       num_input_channels,
+                       input_wav_file.get(),
+                       input_pcm_file.get());
       }
 
       if (msg.has_output_data()) {
-        static FILE* output_file = OpenFile(FLAGS_output_file, "wb");
-        WriteData(msg.output_data().data(), msg.output_data().size(),
-                  output_file, FLAGS_output_file);
-
+        WriteIntData(reinterpret_cast<const int16_t*>(msg.output_data().data()),
+                     msg.output_data().size() / sizeof(int16_t),
+                     output_wav_file.get(),
+                     output_pcm_file.get());
       } else if (msg.output_channel_size() > 0) {
-        static FILE* float_output_file = OpenFile(FLAGS_float_output_file,
-                                                  "wb");
-        // TODO(ajm): Interleave multiple channels.
-        assert(msg.output_channel_size() == 1);
-        WriteData(msg.output_channel(0).data(), msg.output_channel(0).size(),
-                  float_output_file, FLAGS_float_output_file);
+        scoped_ptr<const float*[]> data(new const float*[num_output_channels]);
+        for (int i = 0; i < num_output_channels; ++i) {
+          data[i] =
+              reinterpret_cast<const float*>(msg.output_channel(i).data());
+        }
+        WriteFloatData(data.get(),
+                       msg.output_channel(0).size() / sizeof(float),
+                       num_output_channels,
+                       output_wav_file.get(),
+                       output_pcm_file.get());
       }
 
       if (FLAGS_full) {
@@ -164,15 +252,51 @@
       const Init msg = event_msg.init();
       // These should print out zeros if they're missing.
       fprintf(settings_file, "Init at frame: %d\n", frame_count);
-      fprintf(settings_file, "  Sample rate: %d\n", msg.sample_rate());
-      fprintf(settings_file, "  Input channels: %d\n",
-              msg.num_input_channels());
-      fprintf(settings_file, "  Output channels: %d\n",
-              msg.num_output_channels());
-      fprintf(settings_file, "  Reverse channels: %d\n",
-              msg.num_reverse_channels());
+      int input_sample_rate = msg.sample_rate();
+      fprintf(settings_file, "  Input sample rate: %d\n", input_sample_rate);
+      int output_sample_rate = msg.output_sample_rate();
+      fprintf(settings_file, "  Output sample rate: %d\n", output_sample_rate);
+      int reverse_sample_rate = msg.reverse_sample_rate();
+      fprintf(settings_file,
+              "  Reverse sample rate: %d\n",
+              reverse_sample_rate);
+      num_input_channels = msg.num_input_channels();
+      fprintf(settings_file, "  Input channels: %d\n", num_input_channels);
+      num_output_channels = msg.num_output_channels();
+      fprintf(settings_file, "  Output channels: %d\n", num_output_channels);
+      num_reverse_channels = msg.num_reverse_channels();
+      fprintf(settings_file, "  Reverse channels: %d\n", num_reverse_channels);
 
       fprintf(settings_file, "\n");
+
+      if (reverse_sample_rate == 0) {
+        reverse_sample_rate = input_sample_rate;
+      }
+      if (output_sample_rate == 0) {
+        output_sample_rate = input_sample_rate;
+      }
+
+      if (FLAGS_pcm) {
+        if (!reverse_pcm_file.get()) {
+          reverse_pcm_file.reset(new PcmFile(FLAGS_reverse_file));
+        }
+        if (!input_pcm_file.get()) {
+          input_pcm_file.reset(new PcmFile(FLAGS_input_file));
+        }
+        if (!output_pcm_file.get()) {
+          output_pcm_file.reset(new PcmFile(FLAGS_output_file));
+        }
+      } else {
+        reverse_wav_file.reset(new WavFile(FLAGS_reverse_wav_file,
+                                       reverse_sample_rate,
+                                       num_reverse_channels));
+        input_wav_file.reset(new WavFile(FLAGS_input_wav_file,
+                                     input_sample_rate,
+                                     num_input_channels));
+        output_wav_file.reset(new WavFile(FLAGS_output_wav_file,
+                                      output_sample_rate,
+                                      num_output_channels));
+      }
     }
   }