Revert "Prefix flag macros with WEBRTC_."

This reverts commit 5ccdc1331fcc3cd78eaa14408fe0c38d37a5a51d.

Reason for revert: Breaks downstream project.

Original change's description:
> Prefix flag macros with WEBRTC_.
> 
> Macros defined in rtc_base/flags.h are intended to be used to define
> flags in WebRTC's binaries (e.g. tests).
> 
> They are currently not prefixed and this could cause problems with
> downstream clients since these names are quite common.
> 
> This CL adds the 'WEBRTC_' prefix to them.
> 
> Generated with:
> 
> for x in DECLARE DEFINE; do
>   for y in bool int float string FLAG; do
>     git grep -l "\b$x\_$y\b" | \
>     xargs sed -i "s/\b$x\_$y\b/WEBRTC_$x\_$y/g"
>   done
> done
> git cl format
> 
> Bug: webrtc:9884
> Change-Id: I7b524762b6a3e5aa5b2fc2395edd3e1a0fe72591
> Reviewed-on: https://webrtc-review.googlesource.com/c/106682
> Reviewed-by: Karl Wiberg <kwiberg@webrtc.org>
> Commit-Queue: Mirko Bonadei <mbonadei@webrtc.org>
> Cr-Commit-Position: refs/heads/master@{#25270}

TBR=mbonadei@webrtc.org,kwiberg@webrtc.org

Change-Id: Ia79cd6066ecfd1511c34f1b30fd423e560ed6854
No-Presubmit: true
No-Tree-Checks: true
No-Try: true
Bug: webrtc:9884
Reviewed-on: https://webrtc-review.googlesource.com/c/107160
Reviewed-by: Mirko Bonadei <mbonadei@webrtc.org>
Commit-Queue: Mirko Bonadei <mbonadei@webrtc.org>
Cr-Commit-Position: refs/heads/master@{#25276}
diff --git a/audio/test/low_bandwidth_audio_test.cc b/audio/test/low_bandwidth_audio_test.cc
index 857f983..1009a5e 100644
--- a/audio/test/low_bandwidth_audio_test.cc
+++ b/audio/test/low_bandwidth_audio_test.cc
@@ -14,15 +14,14 @@
 #include "system_wrappers/include/sleep.h"
 #include "test/testsupport/fileutils.h"
 
-WEBRTC_DEFINE_int(sample_rate_hz,
-                  16000,
-                  "Sample rate (Hz) of the produced audio files.");
+DEFINE_int(sample_rate_hz,
+           16000,
+           "Sample rate (Hz) of the produced audio files.");
 
-WEBRTC_DEFINE_bool(
-    quick,
-    false,
-    "Don't do the full audio recording. "
-    "Used to quickly check that the test runs without crashing.");
+DEFINE_bool(quick,
+            false,
+            "Don't do the full audio recording. "
+            "Used to quickly check that the test runs without crashing.");
 
 namespace webrtc {
 namespace test {
diff --git a/call/rampup_tests.cc b/call/rampup_tests.cc
index 6ce17e9..115f968 100644
--- a/call/rampup_tests.cc
+++ b/call/rampup_tests.cc
@@ -40,9 +40,7 @@
 }
 }  // namespace
 
-WEBRTC_DEFINE_string(ramp_dump_name,
-                     "",
-                     "Filename for dumped received RTP stream.");
+DEFINE_string(ramp_dump_name, "", "Filename for dumped received RTP stream.");
 
 RampUpTester::RampUpTester(size_t num_video_streams,
                            size_t num_audio_streams,
diff --git a/examples/peerconnection/client/flagdefs.h b/examples/peerconnection/client/flagdefs.h
index 7bbc383..564e0e9 100644
--- a/examples/peerconnection/client/flagdefs.h
+++ b/examples/peerconnection/client/flagdefs.h
@@ -19,16 +19,16 @@
 // header file so that they can be shared across the different main.cc's
 // for each platform.
 
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
-WEBRTC_DEFINE_bool(autoconnect,
-                   false,
-                   "Connect to the server without user "
-                   "intervention.");
-WEBRTC_DEFINE_string(server, "localhost", "The server to connect to.");
-WEBRTC_DEFINE_int(port,
-                  kDefaultServerPort,
-                  "The port on which the server is listening.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(help, false, "Prints this message");
+DEFINE_bool(autoconnect,
+            false,
+            "Connect to the server without user "
+            "intervention.");
+DEFINE_string(server, "localhost", "The server to connect to.");
+DEFINE_int(port,
+           kDefaultServerPort,
+           "The port on which the server is listening.");
+DEFINE_bool(
     autocall,
     false,
     "Call the first available other client on "
diff --git a/examples/stunprober/main.cc b/examples/stunprober/main.cc
index f438fcf..45a76f5 100644
--- a/examples/stunprober/main.cc
+++ b/examples/stunprober/main.cc
@@ -33,21 +33,16 @@
 using stunprober::StunProber;
 using stunprober::AsyncCallback;
 
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
-WEBRTC_DEFINE_int(interval,
-                  10,
-                  "Interval of consecutive stun pings in milliseconds");
-WEBRTC_DEFINE_bool(shared_socket,
-                   false,
-                   "Share socket mode for different remote IPs");
-WEBRTC_DEFINE_int(pings_per_ip,
-                  10,
-                  "Number of consecutive stun pings to send for each IP");
-WEBRTC_DEFINE_int(
-    timeout,
-    1000,
-    "Milliseconds of wait after the last ping sent before exiting");
-WEBRTC_DEFINE_string(
+DEFINE_bool(help, false, "Prints this message");
+DEFINE_int(interval, 10, "Interval of consecutive stun pings in milliseconds");
+DEFINE_bool(shared_socket, false, "Share socket mode for different remote IPs");
+DEFINE_int(pings_per_ip,
+           10,
+           "Number of consecutive stun pings to send for each IP");
+DEFINE_int(timeout,
+           1000,
+           "Milliseconds of wait after the last ping sent before exiting");
+DEFINE_string(
     servers,
     "stun.l.google.com:19302,stun1.l.google.com:19302,stun2.l.google.com:19302",
     "Comma separated STUN server addresses with ports");
diff --git a/logging/rtc_event_log/rtc_event_log2rtp_dump.cc b/logging/rtc_event_log/rtc_event_log2rtp_dump.cc
index 094af1f..46ce248 100644
--- a/logging/rtc_event_log/rtc_event_log2rtp_dump.cc
+++ b/logging/rtc_event_log/rtc_event_log2rtp_dump.cc
@@ -32,32 +32,31 @@
 
 using MediaType = webrtc::ParsedRtcEventLogNew::MediaType;
 
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     audio,
     true,
     "Use --noaudio to exclude audio packets from the converted RTPdump file.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     video,
     true,
     "Use --novideo to exclude video packets from the converted RTPdump file.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     data,
     true,
     "Use --nodata to exclude data packets from the converted RTPdump file.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     rtp,
     true,
     "Use --nortp to exclude RTP packets from the converted RTPdump file.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     rtcp,
     true,
     "Use --nortcp to exclude RTCP packets from the converted RTPdump file.");
-WEBRTC_DEFINE_string(
-    ssrc,
-    "",
-    "Store only packets with this SSRC (decimal or hex, the latter "
-    "starting with 0x).");
-WEBRTC_DEFINE_bool(help, false, "Prints this message.");
+DEFINE_string(ssrc,
+              "",
+              "Store only packets with this SSRC (decimal or hex, the latter "
+              "starting with 0x).");
+DEFINE_bool(help, false, "Prints this message.");
 
 // Parses the input string for a valid SSRC. If a valid SSRC is found, it is
 // written to the output variable |ssrc|, and true is returned. Otherwise,
diff --git a/logging/rtc_event_log/rtc_event_log2stats.cc b/logging/rtc_event_log/rtc_event_log2stats.cc
index da8f23a..928c829 100644
--- a/logging/rtc_event_log/rtc_event_log2stats.cc
+++ b/logging/rtc_event_log/rtc_event_log2stats.cc
@@ -36,7 +36,7 @@
 
 namespace {
 
-WEBRTC_DEFINE_bool(help, false, "Prints this message.");
+DEFINE_bool(help, false, "Prints this message.");
 
 struct Stats {
   int count = 0;
diff --git a/logging/rtc_event_log/rtc_event_log2text.cc b/logging/rtc_event_log/rtc_event_log2text.cc
index 4cdde7d..67cb6ec 100644
--- a/logging/rtc_event_log/rtc_event_log2text.cc
+++ b/logging/rtc_event_log/rtc_event_log2text.cc
@@ -42,44 +42,35 @@
 
 namespace {
 
-WEBRTC_DEFINE_bool(unknown, true, "Use --nounknown to exclude unknown events.");
-WEBRTC_DEFINE_bool(startstop,
-                   true,
-                   "Use --nostartstop to exclude start/stop events.");
-WEBRTC_DEFINE_bool(config,
-                   true,
-                   "Use --noconfig to exclude stream configurations.");
-WEBRTC_DEFINE_bool(bwe, true, "Use --nobwe to exclude BWE events.");
-WEBRTC_DEFINE_bool(incoming,
-                   true,
-                   "Use --noincoming to exclude incoming packets.");
-WEBRTC_DEFINE_bool(outgoing, true, "Use --nooutgoing to exclude packets.");
+DEFINE_bool(unknown, true, "Use --nounknown to exclude unknown events.");
+DEFINE_bool(startstop, true, "Use --nostartstop to exclude start/stop events.");
+DEFINE_bool(config, true, "Use --noconfig to exclude stream configurations.");
+DEFINE_bool(bwe, true, "Use --nobwe to exclude BWE events.");
+DEFINE_bool(incoming, true, "Use --noincoming to exclude incoming packets.");
+DEFINE_bool(outgoing, true, "Use --nooutgoing to exclude packets.");
 // TODO(terelius): Note that the media type doesn't work with outgoing packets.
-WEBRTC_DEFINE_bool(audio, true, "Use --noaudio to exclude audio packets.");
+DEFINE_bool(audio, true, "Use --noaudio to exclude audio packets.");
 // TODO(terelius): Note that the media type doesn't work with outgoing packets.
-WEBRTC_DEFINE_bool(video, true, "Use --novideo to exclude video packets.");
+DEFINE_bool(video, true, "Use --novideo to exclude video packets.");
 // TODO(terelius): Note that the media type doesn't work with outgoing packets.
-WEBRTC_DEFINE_bool(data, true, "Use --nodata to exclude data packets.");
-WEBRTC_DEFINE_bool(rtp, true, "Use --nortp to exclude RTP packets.");
-WEBRTC_DEFINE_bool(rtcp, true, "Use --nortcp to exclude RTCP packets.");
-WEBRTC_DEFINE_bool(playout,
-                   true,
-                   "Use --noplayout to exclude audio playout events.");
-WEBRTC_DEFINE_bool(ana, true, "Use --noana to exclude ANA events.");
-WEBRTC_DEFINE_bool(probe, true, "Use --noprobe to exclude probe events.");
-WEBRTC_DEFINE_bool(ice, true, "Use --noice to exclude ICE events.");
+DEFINE_bool(data, true, "Use --nodata to exclude data packets.");
+DEFINE_bool(rtp, true, "Use --nortp to exclude RTP packets.");
+DEFINE_bool(rtcp, true, "Use --nortcp to exclude RTCP packets.");
+DEFINE_bool(playout, true, "Use --noplayout to exclude audio playout events.");
+DEFINE_bool(ana, true, "Use --noana to exclude ANA events.");
+DEFINE_bool(probe, true, "Use --noprobe to exclude probe events.");
+DEFINE_bool(ice, true, "Use --noice to exclude ICE events.");
 
-WEBRTC_DEFINE_bool(print_full_packets,
-                   false,
-                   "Print the full RTP headers and RTCP packets in hex.");
+DEFINE_bool(print_full_packets,
+            false,
+            "Print the full RTP headers and RTCP packets in hex.");
 
 // TODO(terelius): Allow a list of SSRCs.
-WEBRTC_DEFINE_string(
-    ssrc,
-    "",
-    "Print only packets with this SSRC (decimal or hex, the latter "
-    "starting with 0x).");
-WEBRTC_DEFINE_bool(help, false, "Prints this message.");
+DEFINE_string(ssrc,
+              "",
+              "Print only packets with this SSRC (decimal or hex, the latter "
+              "starting with 0x).");
+DEFINE_bool(help, false, "Prints this message.");
 
 using MediaType = webrtc::ParsedRtcEventLogNew::MediaType;
 
diff --git a/modules/audio_coding/neteq/neteq_unittest.cc b/modules/audio_coding/neteq/neteq_unittest.cc
index 1c9b9e7..3c63aa7 100644
--- a/modules/audio_coding/neteq/neteq_unittest.cc
+++ b/modules/audio_coding/neteq/neteq_unittest.cc
@@ -52,7 +52,7 @@
 RTC_POP_IGNORING_WUNDEF()
 #endif
 
-WEBRTC_DEFINE_bool(gen_ref, false, "Generate reference files.");
+DEFINE_bool(gen_ref, false, "Generate reference files.");
 
 namespace webrtc {
 
diff --git a/modules/audio_coding/neteq/test/neteq_ilbc_quality_test.cc b/modules/audio_coding/neteq/test/neteq_ilbc_quality_test.cc
index 6f10345..ad61235 100644
--- a/modules/audio_coding/neteq/test/neteq_ilbc_quality_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_ilbc_quality_test.cc
@@ -25,7 +25,7 @@
 static const int kInputSampleRateKhz = 8;
 static const int kOutputSampleRateKhz = 8;
 
-WEBRTC_DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
+DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
 
 }  // namespace
 
diff --git a/modules/audio_coding/neteq/test/neteq_isac_quality_test.cc b/modules/audio_coding/neteq/test/neteq_isac_quality_test.cc
index 651b0ca..94984b87 100644
--- a/modules/audio_coding/neteq/test/neteq_isac_quality_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_isac_quality_test.cc
@@ -21,7 +21,7 @@
 static const int kIsacInputSamplingKhz = 16;
 static const int kIsacOutputSamplingKhz = 16;
 
-WEBRTC_DEFINE_int(bit_rate_kbps, 32, "Target bit rate (kbps).");
+DEFINE_int(bit_rate_kbps, 32, "Target bit rate (kbps).");
 
 }  // namespace
 
diff --git a/modules/audio_coding/neteq/test/neteq_opus_quality_test.cc b/modules/audio_coding/neteq/test/neteq_opus_quality_test.cc
index f4a3636..6861e4c 100644
--- a/modules/audio_coding/neteq/test/neteq_opus_quality_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_opus_quality_test.cc
@@ -22,26 +22,24 @@
 static const int kOpusBlockDurationMs = 20;
 static const int kOpusSamplingKhz = 48;
 
-WEBRTC_DEFINE_int(bit_rate_kbps, 32, "Target bit rate (kbps).");
+DEFINE_int(bit_rate_kbps, 32, "Target bit rate (kbps).");
 
-WEBRTC_DEFINE_int(complexity,
-                  10,
-                  "Complexity: 0 ~ 10 -- defined as in Opus"
-                  "specification.");
+DEFINE_int(complexity,
+           10,
+           "Complexity: 0 ~ 10 -- defined as in Opus"
+           "specification.");
 
-WEBRTC_DEFINE_int(maxplaybackrate, 48000, "Maximum playback rate (Hz).");
+DEFINE_int(maxplaybackrate, 48000, "Maximum playback rate (Hz).");
 
-WEBRTC_DEFINE_int(application, 0, "Application mode: 0 -- VOIP, 1 -- Audio.");
+DEFINE_int(application, 0, "Application mode: 0 -- VOIP, 1 -- Audio.");
 
-WEBRTC_DEFINE_int(reported_loss_rate,
-                  10,
-                  "Reported percentile of packet loss.");
+DEFINE_int(reported_loss_rate, 10, "Reported percentile of packet loss.");
 
-WEBRTC_DEFINE_bool(fec, false, "Enable FEC for encoding (-nofec to disable).");
+DEFINE_bool(fec, false, "Enable FEC for encoding (-nofec to disable).");
 
-WEBRTC_DEFINE_bool(dtx, false, "Enable DTX for encoding (-nodtx to disable).");
+DEFINE_bool(dtx, false, "Enable DTX for encoding (-nodtx to disable).");
 
-WEBRTC_DEFINE_int(sub_packets, 1, "Number of sub packets to repacketize.");
+DEFINE_int(sub_packets, 1, "Number of sub packets to repacketize.");
 
 }  // namespace
 
diff --git a/modules/audio_coding/neteq/test/neteq_pcm16b_quality_test.cc b/modules/audio_coding/neteq/test/neteq_pcm16b_quality_test.cc
index 9c53919..8872b94 100644
--- a/modules/audio_coding/neteq/test/neteq_pcm16b_quality_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_pcm16b_quality_test.cc
@@ -26,7 +26,7 @@
 static const int kInputSampleRateKhz = 48;
 static const int kOutputSampleRateKhz = 48;
 
-WEBRTC_DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
+DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
 
 }  // namespace
 
diff --git a/modules/audio_coding/neteq/test/neteq_pcmu_quality_test.cc b/modules/audio_coding/neteq/test/neteq_pcmu_quality_test.cc
index 85f2267..54ff849 100644
--- a/modules/audio_coding/neteq/test/neteq_pcmu_quality_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_pcmu_quality_test.cc
@@ -25,7 +25,7 @@
 static const int kInputSampleRateKhz = 8;
 static const int kOutputSampleRateKhz = 8;
 
-WEBRTC_DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
+DEFINE_int(frame_size_ms, 20, "Codec frame size (milliseconds).");
 
 }  // namespace
 
diff --git a/modules/audio_coding/neteq/test/neteq_speed_test.cc b/modules/audio_coding/neteq/test/neteq_speed_test.cc
index 70777a2..c1d78c5 100644
--- a/modules/audio_coding/neteq/test/neteq_speed_test.cc
+++ b/modules/audio_coding/neteq/test/neteq_speed_test.cc
@@ -16,10 +16,10 @@
 #include "rtc_base/flags.h"
 
 // Define command line flags.
-WEBRTC_DEFINE_int(runtime_ms, 10000, "Simulated runtime in ms.");
-WEBRTC_DEFINE_int(lossrate, 10, "Packet lossrate; drop every N packets.");
-WEBRTC_DEFINE_float(drift, 0.1f, "Clockdrift factor.");
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_int(runtime_ms, 10000, "Simulated runtime in ms.");
+DEFINE_int(lossrate, 10, "Packet lossrate; drop every N packets.");
+DEFINE_float(drift, 0.1f, "Clockdrift factor.");
+DEFINE_bool(help, false, "Print this message.");
 
 int main(int argc, char* argv[]) {
   std::string program_name = argv[0];
diff --git a/modules/audio_coding/neteq/tools/neteq_quality_test.cc b/modules/audio_coding/neteq/tools/neteq_quality_test.cc
index 2ee6779..faca895 100644
--- a/modules/audio_coding/neteq/tools/neteq_quality_test.cc
+++ b/modules/audio_coding/neteq/tools/neteq_quality_test.cc
@@ -47,47 +47,42 @@
   return true;
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     in_filename,
     DefaultInFilename().c_str(),
     "Filename for input audio (specify sample rate with --input_sample_rate, "
     "and channels with --channels).");
 
-WEBRTC_DEFINE_int(input_sample_rate, 16000, "Sample rate of input file in Hz.");
+DEFINE_int(input_sample_rate, 16000, "Sample rate of input file in Hz.");
 
-WEBRTC_DEFINE_int(channels, 1, "Number of channels in input audio.");
+DEFINE_int(channels, 1, "Number of channels in input audio.");
 
-WEBRTC_DEFINE_string(out_filename,
-                     DefaultOutFilename().c_str(),
-                     "Name of output audio file.");
+DEFINE_string(out_filename,
+              DefaultOutFilename().c_str(),
+              "Name of output audio file.");
 
-WEBRTC_DEFINE_int(runtime_ms, 10000, "Simulated runtime (milliseconds).");
+DEFINE_int(runtime_ms, 10000, "Simulated runtime (milliseconds).");
 
-WEBRTC_DEFINE_int(packet_loss_rate, 10, "Percentile of packet loss.");
+DEFINE_int(packet_loss_rate, 10, "Percentile of packet loss.");
 
-WEBRTC_DEFINE_int(
-    random_loss_mode,
-    kUniformLoss,
-    "Random loss mode: 0--no loss, 1--uniform loss, 2--Gilbert Elliot "
-    "loss, 3--fixed loss.");
+DEFINE_int(random_loss_mode,
+           kUniformLoss,
+           "Random loss mode: 0--no loss, 1--uniform loss, 2--Gilbert Elliot "
+           "loss, 3--fixed loss.");
 
-WEBRTC_DEFINE_int(
-    burst_length,
-    30,
-    "Burst length in milliseconds, only valid for Gilbert Elliot loss.");
+DEFINE_int(burst_length,
+           30,
+           "Burst length in milliseconds, only valid for Gilbert Elliot loss.");
 
-WEBRTC_DEFINE_float(drift_factor, 0.0, "Time drift factor.");
+DEFINE_float(drift_factor, 0.0, "Time drift factor.");
 
-WEBRTC_DEFINE_int(preload_packets,
-                  0,
-                  "Preload the buffer with this many packets.");
+DEFINE_int(preload_packets, 0, "Preload the buffer with this many packets.");
 
-WEBRTC_DEFINE_string(
-    loss_events,
-    "",
-    "List of loss events time and duration separated by comma: "
-    "<first_event_time> <first_event_duration>, <second_event_time> "
-    "<second_event_duration>, ...");
+DEFINE_string(loss_events,
+              "",
+              "List of loss events time and duration separated by comma: "
+              "<first_event_time> <first_event_duration>, <second_event_time> "
+              "<second_event_duration>, ...");
 
 // ProbTrans00Solver() is to calculate the transition probability from no-loss
 // state to itself in a modified Gilbert Elliot packet loss model. The result is
diff --git a/modules/audio_coding/neteq/tools/neteq_rtpplay.cc b/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
index c2726eb..25e8cd8 100644
--- a/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
+++ b/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
@@ -17,17 +17,17 @@
 #include "system_wrappers/include/field_trial.h"
 #include "test/field_trial.h"
 
-WEBRTC_DEFINE_bool(codec_map,
-                   false,
-                   "Prints the mapping between RTP payload type and "
-                   "codec");
-WEBRTC_DEFINE_string(
+DEFINE_bool(codec_map,
+            false,
+            "Prints the mapping between RTP payload type and "
+            "codec");
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
     "E.g. running with --force_fieldtrials=WebRTC-FooFeature/Enable/"
     " will assign the group Enable to field trial WebRTC-FooFeature.");
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
+DEFINE_bool(help, false, "Prints this message");
 
 int main(int argc, char* argv[]) {
   webrtc::test::NetEqTestFactory factory;
diff --git a/modules/audio_coding/neteq/tools/neteq_test_factory.cc b/modules/audio_coding/neteq/tools/neteq_test_factory.cc
index 93da54c..df3a9f0 100644
--- a/modules/audio_coding/neteq/tools/neteq_test_factory.cc
+++ b/modules/audio_coding/neteq/tools/neteq_test_factory.cc
@@ -91,57 +91,50 @@
 }
 
 // Define command line flags.
-WEBRTC_DEFINE_int(pcmu, 0, "RTP payload type for PCM-u");
-WEBRTC_DEFINE_int(pcma, 8, "RTP payload type for PCM-a");
-WEBRTC_DEFINE_int(ilbc, 102, "RTP payload type for iLBC");
-WEBRTC_DEFINE_int(isac, 103, "RTP payload type for iSAC");
-WEBRTC_DEFINE_int(isac_swb, 104, "RTP payload type for iSAC-swb (32 kHz)");
-WEBRTC_DEFINE_int(opus, 111, "RTP payload type for Opus");
-WEBRTC_DEFINE_int(pcm16b, 93, "RTP payload type for PCM16b-nb (8 kHz)");
-WEBRTC_DEFINE_int(pcm16b_wb, 94, "RTP payload type for PCM16b-wb (16 kHz)");
-WEBRTC_DEFINE_int(pcm16b_swb32,
-                  95,
-                  "RTP payload type for PCM16b-swb32 (32 kHz)");
-WEBRTC_DEFINE_int(pcm16b_swb48,
-                  96,
-                  "RTP payload type for PCM16b-swb48 (48 kHz)");
-WEBRTC_DEFINE_int(g722, 9, "RTP payload type for G.722");
-WEBRTC_DEFINE_int(avt, 106, "RTP payload type for AVT/DTMF (8 kHz)");
-WEBRTC_DEFINE_int(avt_16, 114, "RTP payload type for AVT/DTMF (16 kHz)");
-WEBRTC_DEFINE_int(avt_32, 115, "RTP payload type for AVT/DTMF (32 kHz)");
-WEBRTC_DEFINE_int(avt_48, 116, "RTP payload type for AVT/DTMF (48 kHz)");
-WEBRTC_DEFINE_int(red, 117, "RTP payload type for redundant audio (RED)");
-WEBRTC_DEFINE_int(cn_nb, 13, "RTP payload type for comfort noise (8 kHz)");
-WEBRTC_DEFINE_int(cn_wb, 98, "RTP payload type for comfort noise (16 kHz)");
-WEBRTC_DEFINE_int(cn_swb32, 99, "RTP payload type for comfort noise (32 kHz)");
-WEBRTC_DEFINE_int(cn_swb48, 100, "RTP payload type for comfort noise (48 kHz)");
-WEBRTC_DEFINE_string(replacement_audio_file,
-                     "",
-                     "A PCM file that will be used to populate "
-                     "dummy"
-                     " RTP packets");
-WEBRTC_DEFINE_string(
-    ssrc,
-    "",
-    "Only use packets with this SSRC (decimal or hex, the latter "
-    "starting with 0x)");
-WEBRTC_DEFINE_int(audio_level, 1, "Extension ID for audio level (RFC 6464)");
-WEBRTC_DEFINE_int(abs_send_time, 3, "Extension ID for absolute sender time");
-WEBRTC_DEFINE_int(transport_seq_no,
-                  5,
-                  "Extension ID for transport sequence number");
-WEBRTC_DEFINE_int(video_content_type, 7, "Extension ID for video content type");
-WEBRTC_DEFINE_int(video_timing, 8, "Extension ID for video timing");
-WEBRTC_DEFINE_bool(matlabplot,
-                   false,
-                   "Generates a matlab script for plotting the delay profile");
-WEBRTC_DEFINE_bool(pythonplot,
-                   false,
-                   "Generates a python script for plotting the delay profile");
-WEBRTC_DEFINE_bool(concealment_events, false, "Prints concealment events");
-WEBRTC_DEFINE_int(max_nr_packets_in_buffer,
-                  50,
-                  "Maximum allowed number of packets in the buffer");
+DEFINE_int(pcmu, 0, "RTP payload type for PCM-u");
+DEFINE_int(pcma, 8, "RTP payload type for PCM-a");
+DEFINE_int(ilbc, 102, "RTP payload type for iLBC");
+DEFINE_int(isac, 103, "RTP payload type for iSAC");
+DEFINE_int(isac_swb, 104, "RTP payload type for iSAC-swb (32 kHz)");
+DEFINE_int(opus, 111, "RTP payload type for Opus");
+DEFINE_int(pcm16b, 93, "RTP payload type for PCM16b-nb (8 kHz)");
+DEFINE_int(pcm16b_wb, 94, "RTP payload type for PCM16b-wb (16 kHz)");
+DEFINE_int(pcm16b_swb32, 95, "RTP payload type for PCM16b-swb32 (32 kHz)");
+DEFINE_int(pcm16b_swb48, 96, "RTP payload type for PCM16b-swb48 (48 kHz)");
+DEFINE_int(g722, 9, "RTP payload type for G.722");
+DEFINE_int(avt, 106, "RTP payload type for AVT/DTMF (8 kHz)");
+DEFINE_int(avt_16, 114, "RTP payload type for AVT/DTMF (16 kHz)");
+DEFINE_int(avt_32, 115, "RTP payload type for AVT/DTMF (32 kHz)");
+DEFINE_int(avt_48, 116, "RTP payload type for AVT/DTMF (48 kHz)");
+DEFINE_int(red, 117, "RTP payload type for redundant audio (RED)");
+DEFINE_int(cn_nb, 13, "RTP payload type for comfort noise (8 kHz)");
+DEFINE_int(cn_wb, 98, "RTP payload type for comfort noise (16 kHz)");
+DEFINE_int(cn_swb32, 99, "RTP payload type for comfort noise (32 kHz)");
+DEFINE_int(cn_swb48, 100, "RTP payload type for comfort noise (48 kHz)");
+DEFINE_string(replacement_audio_file,
+              "",
+              "A PCM file that will be used to populate "
+              "dummy"
+              " RTP packets");
+DEFINE_string(ssrc,
+              "",
+              "Only use packets with this SSRC (decimal or hex, the latter "
+              "starting with 0x)");
+DEFINE_int(audio_level, 1, "Extension ID for audio level (RFC 6464)");
+DEFINE_int(abs_send_time, 3, "Extension ID for absolute sender time");
+DEFINE_int(transport_seq_no, 5, "Extension ID for transport sequence number");
+DEFINE_int(video_content_type, 7, "Extension ID for video content type");
+DEFINE_int(video_timing, 8, "Extension ID for video timing");
+DEFINE_bool(matlabplot,
+            false,
+            "Generates a matlab script for plotting the delay profile");
+DEFINE_bool(pythonplot,
+            false,
+            "Generates a python script for plotting the delay profile");
+DEFINE_bool(concealment_events, false, "Prints concealment events");
+DEFINE_int(max_nr_packets_in_buffer,
+           50,
+           "Maximum allowed number of packets in the buffer");
 
 // Maps a codec type to a printable name string.
 std::string CodecName(NetEqDecoder codec) {
diff --git a/modules/audio_coding/neteq/tools/rtp_analyze.cc b/modules/audio_coding/neteq/tools/rtp_analyze.cc
index 9d3041e..f939038 100644
--- a/modules/audio_coding/neteq/tools/rtp_analyze.cc
+++ b/modules/audio_coding/neteq/tools/rtp_analyze.cc
@@ -19,16 +19,16 @@
 #include "rtc_base/flags.h"
 
 // Define command line flags.
-WEBRTC_DEFINE_int(red, 117, "RTP payload type for RED");
-WEBRTC_DEFINE_int(audio_level,
-                  -1,
-                  "Extension ID for audio level (RFC 6464); "
-                  "-1 not to print audio level");
-WEBRTC_DEFINE_int(abs_send_time,
-                  -1,
-                  "Extension ID for absolute sender time; "
-                  "-1 not to print absolute send time");
-WEBRTC_DEFINE_bool(help, false, "Print this message");
+DEFINE_int(red, 117, "RTP payload type for RED");
+DEFINE_int(audio_level,
+           -1,
+           "Extension ID for audio level (RFC 6464); "
+           "-1 not to print audio level");
+DEFINE_int(abs_send_time,
+           -1,
+           "Extension ID for absolute sender time; "
+           "-1 not to print absolute send time");
+DEFINE_bool(help, false, "Print this message");
 
 int main(int argc, char* argv[]) {
   std::string program_name = argv[0];
diff --git a/modules/audio_coding/neteq/tools/rtp_encode.cc b/modules/audio_coding/neteq/tools/rtp_encode.cc
index f48b04d..5065ca1 100644
--- a/modules/audio_coding/neteq/tools/rtp_encode.cc
+++ b/modules/audio_coding/neteq/tools/rtp_encode.cc
@@ -40,24 +40,20 @@
 namespace {
 
 // Define command line flags.
-WEBRTC_DEFINE_bool(list_codecs, false, "Enumerate all codecs");
-WEBRTC_DEFINE_string(codec, "opus", "Codec to use");
-WEBRTC_DEFINE_int(frame_len,
-                  0,
-                  "Frame length in ms; 0 indicates codec default value");
-WEBRTC_DEFINE_int(bitrate,
-                  0,
-                  "Bitrate in kbps; 0 indicates codec default value");
-WEBRTC_DEFINE_int(payload_type,
-                  -1,
-                  "RTP payload type; -1 indicates codec default value");
-WEBRTC_DEFINE_int(cng_payload_type,
-                  -1,
-                  "RTP payload type for CNG; -1 indicates default value");
-WEBRTC_DEFINE_int(ssrc, 0, "SSRC to write to the RTP header");
-WEBRTC_DEFINE_bool(dtx, false, "Use DTX/CNG");
-WEBRTC_DEFINE_int(sample_rate, 48000, "Sample rate of the input file");
-WEBRTC_DEFINE_bool(help, false, "Print this message");
+DEFINE_bool(list_codecs, false, "Enumerate all codecs");
+DEFINE_string(codec, "opus", "Codec to use");
+DEFINE_int(frame_len, 0, "Frame length in ms; 0 indicates codec default value");
+DEFINE_int(bitrate, 0, "Bitrate in kbps; 0 indicates codec default value");
+DEFINE_int(payload_type,
+           -1,
+           "RTP payload type; -1 indicates codec default value");
+DEFINE_int(cng_payload_type,
+           -1,
+           "RTP payload type for CNG; -1 indicates default value");
+DEFINE_int(ssrc, 0, "SSRC to write to the RTP header");
+DEFINE_bool(dtx, false, "Use DTX/CNG");
+DEFINE_int(sample_rate, 48000, "Sample rate of the input file");
+DEFINE_bool(help, false, "Print this message");
 
 // Add new codecs here, and to the map below.
 enum class CodecType {
diff --git a/modules/audio_coding/neteq/tools/rtp_jitter.cc b/modules/audio_coding/neteq/tools/rtp_jitter.cc
index 92a7a8d..3c49443 100644
--- a/modules/audio_coding/neteq/tools/rtp_jitter.cc
+++ b/modules/audio_coding/neteq/tools/rtp_jitter.cc
@@ -23,7 +23,7 @@
 namespace test {
 namespace {
 
-WEBRTC_DEFINE_bool(help, false, "Print help message");
+DEFINE_bool(help, false, "Print help message");
 
 constexpr size_t kRtpDumpHeaderLength = 8;
 
diff --git a/modules/audio_mixer/audio_mixer_test.cc b/modules/audio_mixer/audio_mixer_test.cc
index b607877..2004aee 100644
--- a/modules/audio_mixer/audio_mixer_test.cc
+++ b/modules/audio_mixer/audio_mixer_test.cc
@@ -19,25 +19,24 @@
 #include "rtc_base/flags.h"
 #include "rtc_base/strings/string_builder.h"
 
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
-WEBRTC_DEFINE_int(
-    sampling_rate,
-    16000,
-    "Rate at which to mix (all input streams must have this rate)");
+DEFINE_bool(help, false, "Prints this message");
+DEFINE_int(sampling_rate,
+           16000,
+           "Rate at which to mix (all input streams must have this rate)");
 
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     stereo,
     false,
     "Enable stereo (interleaved). Inputs need not be as this parameter.");
 
-WEBRTC_DEFINE_bool(limiter, true, "Enable limiter.");
-WEBRTC_DEFINE_string(output_file,
-                     "mixed_file.wav",
-                     "File in which to store the mixed result.");
-WEBRTC_DEFINE_string(input_file_1, "", "First input. Default none.");
-WEBRTC_DEFINE_string(input_file_2, "", "Second input. Default none.");
-WEBRTC_DEFINE_string(input_file_3, "", "Third input. Default none.");
-WEBRTC_DEFINE_string(input_file_4, "", "Fourth input. Default none.");
+DEFINE_bool(limiter, true, "Enable limiter.");
+DEFINE_string(output_file,
+              "mixed_file.wav",
+              "File in which to store the mixed result.");
+DEFINE_string(input_file_1, "", "First input. Default none.");
+DEFINE_string(input_file_2, "", "Second input. Default none.");
+DEFINE_string(input_file_3, "", "Third input. Default none.");
+DEFINE_string(input_file_4, "", "Fourth input. Default none.");
 
 namespace webrtc {
 namespace test {
diff --git a/modules/audio_processing/agc2/rnn_vad/rnn_vad_tool.cc b/modules/audio_processing/agc2/rnn_vad/rnn_vad_tool.cc
index b66dfd6..5fba0bf 100644
--- a/modules/audio_processing/agc2/rnn_vad/rnn_vad_tool.cc
+++ b/modules/audio_processing/agc2/rnn_vad/rnn_vad_tool.cc
@@ -25,22 +25,22 @@
 namespace test {
 namespace {
 
-WEBRTC_DEFINE_string(i, "", "Path to the input wav file");
+DEFINE_string(i, "", "Path to the input wav file");
 std::string InputWavFile() {
   return static_cast<std::string>(FLAG_i);
 }
 
-WEBRTC_DEFINE_string(f, "", "Path to the output features file");
+DEFINE_string(f, "", "Path to the output features file");
 std::string OutputFeaturesFile() {
   return static_cast<std::string>(FLAG_f);
 }
 
-WEBRTC_DEFINE_string(o, "", "Path to the output VAD probabilities file");
+DEFINE_string(o, "", "Path to the output VAD probabilities file");
 std::string OutputVadProbsFile() {
   return static_cast<std::string>(FLAG_o);
 }
 
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
+DEFINE_bool(help, false, "Prints this message");
 
 }  // namespace
 
diff --git a/modules/audio_processing/test/audioproc_float_impl.cc b/modules/audio_processing/test/audioproc_float_impl.cc
index e3f5004..3a44887 100644
--- a/modules/audio_processing/test/audioproc_float_impl.cc
+++ b/modules/audio_processing/test/audioproc_float_impl.cc
@@ -37,170 +37,155 @@
     "processing module, either based on wav files or "
     "protobuf debug dump recordings.\n";
 
-WEBRTC_DEFINE_string(dump_input, "", "Aec dump input filename");
-WEBRTC_DEFINE_string(dump_output, "", "Aec dump output filename");
-WEBRTC_DEFINE_string(i, "", "Forward stream input wav filename");
-WEBRTC_DEFINE_string(o, "", "Forward stream output wav filename");
-WEBRTC_DEFINE_string(ri, "", "Reverse stream input wav filename");
-WEBRTC_DEFINE_string(ro, "", "Reverse stream output wav filename");
-WEBRTC_DEFINE_string(artificial_nearend, "", "Artificial nearend wav filename");
-WEBRTC_DEFINE_int(output_num_channels,
-                  kParameterNotSpecifiedValue,
-                  "Number of forward stream output channels");
-WEBRTC_DEFINE_int(reverse_output_num_channels,
-                  kParameterNotSpecifiedValue,
-                  "Number of Reverse stream output channels");
-WEBRTC_DEFINE_int(output_sample_rate_hz,
-                  kParameterNotSpecifiedValue,
-                  "Forward stream output sample rate in Hz");
-WEBRTC_DEFINE_int(reverse_output_sample_rate_hz,
-                  kParameterNotSpecifiedValue,
-                  "Reverse stream output sample rate in Hz");
-WEBRTC_DEFINE_bool(fixed_interface,
-                   false,
-                   "Use the fixed interface when operating on wav files");
-WEBRTC_DEFINE_int(aec,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the echo canceller");
-WEBRTC_DEFINE_int(aecm,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the mobile echo controller");
-WEBRTC_DEFINE_int(ed,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate (0) the residual echo detector");
-WEBRTC_DEFINE_string(ed_graph,
-                     "",
-                     "Output filename for graph of echo likelihood");
-WEBRTC_DEFINE_int(agc,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the AGC");
-WEBRTC_DEFINE_int(agc2,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the AGC2");
-WEBRTC_DEFINE_int(pre_amplifier,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the pre amplifier");
-WEBRTC_DEFINE_int(hpf,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the high-pass filter");
-WEBRTC_DEFINE_int(ns,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the noise suppressor");
-WEBRTC_DEFINE_int(ts,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the transient suppressor");
-WEBRTC_DEFINE_int(vad,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the voice activity detector");
-WEBRTC_DEFINE_int(le,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the level estimator");
-WEBRTC_DEFINE_bool(
-    all_default,
-    false,
-    "Activate all of the default components (will be overridden by any "
-    "other settings)");
-WEBRTC_DEFINE_int(aec_suppression_level,
-                  kParameterNotSpecifiedValue,
-                  "Set the aec suppression level (0-2)");
-WEBRTC_DEFINE_int(delay_agnostic,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the AEC delay agnostic mode");
-WEBRTC_DEFINE_int(extended_filter,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the AEC extended filter mode");
-WEBRTC_DEFINE_int(
-    aec3,
-    kParameterNotSpecifiedValue,
-    "Activate (1) or deactivate(0) the experimental AEC mode AEC3");
-WEBRTC_DEFINE_int(experimental_agc,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the experimental AGC");
-WEBRTC_DEFINE_int(
-    experimental_agc_disable_digital_adaptive,
-    kParameterNotSpecifiedValue,
-    "Force-deactivate (1) digital adaptation in "
-    "experimental AGC. Digital adaptation is active by default (0).");
-WEBRTC_DEFINE_int(experimental_agc_analyze_before_aec,
-                  kParameterNotSpecifiedValue,
-                  "Make level estimation happen before AEC"
-                  " in the experimental AGC. After AEC is the default (0)");
-WEBRTC_DEFINE_int(
+DEFINE_string(dump_input, "", "Aec dump input filename");
+DEFINE_string(dump_output, "", "Aec dump output filename");
+DEFINE_string(i, "", "Forward stream input wav filename");
+DEFINE_string(o, "", "Forward stream output wav filename");
+DEFINE_string(ri, "", "Reverse stream input wav filename");
+DEFINE_string(ro, "", "Reverse stream output wav filename");
+DEFINE_string(artificial_nearend, "", "Artificial nearend wav filename");
+DEFINE_int(output_num_channels,
+           kParameterNotSpecifiedValue,
+           "Number of forward stream output channels");
+DEFINE_int(reverse_output_num_channels,
+           kParameterNotSpecifiedValue,
+           "Number of Reverse stream output channels");
+DEFINE_int(output_sample_rate_hz,
+           kParameterNotSpecifiedValue,
+           "Forward stream output sample rate in Hz");
+DEFINE_int(reverse_output_sample_rate_hz,
+           kParameterNotSpecifiedValue,
+           "Reverse stream output sample rate in Hz");
+DEFINE_bool(fixed_interface,
+            false,
+            "Use the fixed interface when operating on wav files");
+DEFINE_int(aec,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the echo canceller");
+DEFINE_int(aecm,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the mobile echo controller");
+DEFINE_int(ed,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate (0) the residual echo detector");
+DEFINE_string(ed_graph, "", "Output filename for graph of echo likelihood");
+DEFINE_int(agc,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the AGC");
+DEFINE_int(agc2,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the AGC2");
+DEFINE_int(pre_amplifier,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the pre amplifier");
+DEFINE_int(hpf,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the high-pass filter");
+DEFINE_int(ns,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the noise suppressor");
+DEFINE_int(ts,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the transient suppressor");
+DEFINE_int(vad,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the voice activity detector");
+DEFINE_int(le,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the level estimator");
+DEFINE_bool(all_default,
+            false,
+            "Activate all of the default components (will be overridden by any "
+            "other settings)");
+DEFINE_int(aec_suppression_level,
+           kParameterNotSpecifiedValue,
+           "Set the aec suppression level (0-2)");
+DEFINE_int(delay_agnostic,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the AEC delay agnostic mode");
+DEFINE_int(extended_filter,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the AEC extended filter mode");
+DEFINE_int(aec3,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the experimental AEC mode AEC3");
+DEFINE_int(experimental_agc,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the experimental AGC");
+DEFINE_int(experimental_agc_disable_digital_adaptive,
+           kParameterNotSpecifiedValue,
+           "Force-deactivate (1) digital adaptation in "
+           "experimental AGC. Digital adaptation is active by default (0).");
+DEFINE_int(experimental_agc_analyze_before_aec,
+           kParameterNotSpecifiedValue,
+           "Make level estimation happen before AEC"
+           " in the experimental AGC. After AEC is the default (0)");
+DEFINE_int(
     experimental_agc_agc2_level_estimator,
     kParameterNotSpecifiedValue,
     "AGC2 level estimation"
     " in the experimental AGC. AGC1 level estimation is the default (0)");
-WEBRTC_DEFINE_int(
+DEFINE_int(
     refined_adaptive_filter,
     kParameterNotSpecifiedValue,
     "Activate (1) or deactivate(0) the refined adaptive filter functionality");
-WEBRTC_DEFINE_int(agc_mode,
-                  kParameterNotSpecifiedValue,
-                  "Specify the AGC mode (0-2)");
-WEBRTC_DEFINE_int(agc_target_level,
-                  kParameterNotSpecifiedValue,
-                  "Specify the AGC target level (0-31)");
-WEBRTC_DEFINE_int(agc_limiter,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) the level estimator");
-WEBRTC_DEFINE_int(agc_compression_gain,
-                  kParameterNotSpecifiedValue,
-                  "Specify the AGC compression gain (0-90)");
-WEBRTC_DEFINE_float(agc2_enable_adaptive_gain,
-                    kParameterNotSpecifiedValue,
-                    "Activate (1) or deactivate(0) the AGC2 adaptive gain");
-WEBRTC_DEFINE_float(agc2_fixed_gain_db, 0.f, "AGC2 fixed gain (dB) to apply");
-WEBRTC_DEFINE_float(pre_amplifier_gain_factor,
-                    1.f,
-                    "Pre-amplifier gain factor (linear) to apply");
-WEBRTC_DEFINE_int(vad_likelihood,
-                  kParameterNotSpecifiedValue,
-                  "Specify the VAD likelihood (0-3)");
-WEBRTC_DEFINE_int(ns_level,
-                  kParameterNotSpecifiedValue,
-                  "Specify the NS level (0-3)");
-WEBRTC_DEFINE_int(stream_delay,
-                  kParameterNotSpecifiedValue,
-                  "Specify the stream delay in ms to use");
-WEBRTC_DEFINE_int(use_stream_delay,
-                  kParameterNotSpecifiedValue,
-                  "Activate (1) or deactivate(0) reporting the stream delay");
-WEBRTC_DEFINE_int(stream_drift_samples,
-                  kParameterNotSpecifiedValue,
-                  "Specify the number of stream drift samples to use");
-WEBRTC_DEFINE_int(initial_mic_level, 100, "Initial mic level (0-255)");
-WEBRTC_DEFINE_int(
-    simulate_mic_gain,
-    0,
-    "Activate (1) or deactivate(0) the analog mic gain simulation");
-WEBRTC_DEFINE_int(
-    simulated_mic_kind,
-    kParameterNotSpecifiedValue,
-    "Specify which microphone kind to use for microphone simulation");
-WEBRTC_DEFINE_bool(performance_report, false, "Report the APM performance ");
-WEBRTC_DEFINE_bool(verbose, false, "Produce verbose output");
-WEBRTC_DEFINE_bool(quiet,
-                   false,
-                   "Avoid producing information about the progress.");
-WEBRTC_DEFINE_bool(bitexactness_report,
-                   false,
-                   "Report bitexactness for aec dump result reproduction");
-WEBRTC_DEFINE_bool(discard_settings_in_aecdump,
-                   false,
-                   "Discard any config settings specified in the aec dump");
-WEBRTC_DEFINE_bool(store_intermediate_output,
-                   false,
-                   "Creates new output files after each init");
-WEBRTC_DEFINE_string(custom_call_order_file,
-                     "",
-                     "Custom process API call order file");
-WEBRTC_DEFINE_bool(print_aec3_parameter_values,
-                   false,
-                   "Print parameter values used in AEC3 in JSON-format");
-WEBRTC_DEFINE_string(aec3_settings,
-                     "",
-                     "File in JSON-format with custom AEC3 settings");
-WEBRTC_DEFINE_bool(help, false, "Print this message");
+DEFINE_int(agc_mode, kParameterNotSpecifiedValue, "Specify the AGC mode (0-2)");
+DEFINE_int(agc_target_level,
+           kParameterNotSpecifiedValue,
+           "Specify the AGC target level (0-31)");
+DEFINE_int(agc_limiter,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) the level estimator");
+DEFINE_int(agc_compression_gain,
+           kParameterNotSpecifiedValue,
+           "Specify the AGC compression gain (0-90)");
+DEFINE_float(agc2_enable_adaptive_gain,
+             kParameterNotSpecifiedValue,
+             "Activate (1) or deactivate(0) the AGC2 adaptive gain");
+DEFINE_float(agc2_fixed_gain_db, 0.f, "AGC2 fixed gain (dB) to apply");
+DEFINE_float(pre_amplifier_gain_factor,
+             1.f,
+             "Pre-amplifier gain factor (linear) to apply");
+DEFINE_int(vad_likelihood,
+           kParameterNotSpecifiedValue,
+           "Specify the VAD likelihood (0-3)");
+DEFINE_int(ns_level, kParameterNotSpecifiedValue, "Specify the NS level (0-3)");
+DEFINE_int(stream_delay,
+           kParameterNotSpecifiedValue,
+           "Specify the stream delay in ms to use");
+DEFINE_int(use_stream_delay,
+           kParameterNotSpecifiedValue,
+           "Activate (1) or deactivate(0) reporting the stream delay");
+DEFINE_int(stream_drift_samples,
+           kParameterNotSpecifiedValue,
+           "Specify the number of stream drift samples to use");
+DEFINE_int(initial_mic_level, 100, "Initial mic level (0-255)");
+DEFINE_int(simulate_mic_gain,
+           0,
+           "Activate (1) or deactivate(0) the analog mic gain simulation");
+DEFINE_int(simulated_mic_kind,
+           kParameterNotSpecifiedValue,
+           "Specify which microphone kind to use for microphone simulation");
+DEFINE_bool(performance_report, false, "Report the APM performance ");
+DEFINE_bool(verbose, false, "Produce verbose output");
+DEFINE_bool(quiet, false, "Avoid producing information about the progress.");
+DEFINE_bool(bitexactness_report,
+            false,
+            "Report bitexactness for aec dump result reproduction");
+DEFINE_bool(discard_settings_in_aecdump,
+            false,
+            "Discard any config settings specified in the aec dump");
+DEFINE_bool(store_intermediate_output,
+            false,
+            "Creates new output files after each init");
+DEFINE_string(custom_call_order_file, "", "Custom process API call order file");
+DEFINE_bool(print_aec3_parameter_values,
+            false,
+            "Print parameter values used in AEC3 in JSON-format");
+DEFINE_string(aec3_settings,
+              "",
+              "File in JSON-format with custom AEC3 settings");
+DEFINE_bool(help, false, "Print this message");
 
 void SetSettingIfSpecified(const std::string& value,
                            absl::optional<std::string>* parameter) {
diff --git a/modules/audio_processing/test/conversational_speech/generator.cc b/modules/audio_processing/test/conversational_speech/generator.cc
index fa561cf..741a1ca 100644
--- a/modules/audio_processing/test/conversational_speech/generator.cc
+++ b/modules/audio_processing/test/conversational_speech/generator.cc
@@ -32,10 +32,10 @@
     "Command-line tool to generate multiple-end audio tracks to simulate "
     "conversational speech with two or more participants.\n";
 
-WEBRTC_DEFINE_string(i, "", "Directory containing the speech turn wav files");
-WEBRTC_DEFINE_string(t, "", "Path to the timing text file");
-WEBRTC_DEFINE_string(o, "", "Output wav files destination path");
-WEBRTC_DEFINE_bool(help, false, "Prints this message");
+DEFINE_string(i, "", "Directory containing the speech turn wav files");
+DEFINE_string(t, "", "Path to the timing text file");
+DEFINE_string(o, "", "Output wav files destination path");
+DEFINE_bool(help, false, "Prints this message");
 
 }  // namespace
 
diff --git a/modules/audio_processing/test/py_quality_assessment/quality_assessment/apm_vad.cc b/modules/audio_processing/test/py_quality_assessment/quality_assessment/apm_vad.cc
index 4b6ada2..a6184b5 100644
--- a/modules/audio_processing/test/py_quality_assessment/quality_assessment/apm_vad.cc
+++ b/modules/audio_processing/test/py_quality_assessment/quality_assessment/apm_vad.cc
@@ -24,9 +24,9 @@
 constexpr size_t kMaxFrameLen =
     kAudioFrameLengthMilliseconds * kMaxSampleRate / 1000;
 
-WEBRTC_DEFINE_string(i, "", "Input wav file");
-WEBRTC_DEFINE_string(o_probs, "", "VAD probabilities output file");
-WEBRTC_DEFINE_string(o_rms, "", "VAD output file");
+DEFINE_string(i, "", "Input wav file");
+DEFINE_string(o_probs, "", "VAD probabilities output file");
+DEFINE_string(o_rms, "", "VAD output file");
 
 int main(int argc, char* argv[]) {
   if (rtc::FlagList::SetFlagsFromCommandLine(&argc, argv, true))
diff --git a/modules/audio_processing/test/py_quality_assessment/quality_assessment/sound_level.cc b/modules/audio_processing/test/py_quality_assessment/quality_assessment/sound_level.cc
index 35a2c11..98cf84c 100644
--- a/modules/audio_processing/test/py_quality_assessment/quality_assessment/sound_level.cc
+++ b/modules/audio_processing/test/py_quality_assessment/quality_assessment/sound_level.cc
@@ -26,13 +26,13 @@
 
 const double kOneDbReduction = DbToRatio(-1.0);
 
-WEBRTC_DEFINE_string(i, "", "Input wav file");
-WEBRTC_DEFINE_string(oc, "", "Config output file");
-WEBRTC_DEFINE_string(ol, "", "Levels output file");
-WEBRTC_DEFINE_float(a, 5.f, "Attack (ms)");
-WEBRTC_DEFINE_float(d, 20.f, "Decay (ms)");
-WEBRTC_DEFINE_int(f, 10, "Frame length (ms)");
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_string(i, "", "Input wav file");
+DEFINE_string(oc, "", "Config output file");
+DEFINE_string(ol, "", "Levels output file");
+DEFINE_float(a, 5.f, "Attack (ms)");
+DEFINE_float(d, 20.f, "Decay (ms)");
+DEFINE_int(f, 10, "Frame length (ms)");
+DEFINE_bool(help, false, "prints this message");
 
 int main(int argc, char* argv[]) {
   if (rtc::FlagList::SetFlagsFromCommandLine(&argc, argv, true)) {
diff --git a/modules/audio_processing/test/py_quality_assessment/quality_assessment/vad.cc b/modules/audio_processing/test/py_quality_assessment/quality_assessment/vad.cc
index 8a134ed..191cb1e 100644
--- a/modules/audio_processing/test/py_quality_assessment/quality_assessment/vad.cc
+++ b/modules/audio_processing/test/py_quality_assessment/quality_assessment/vad.cc
@@ -27,8 +27,8 @@
 
 constexpr uint8_t kBitmaskBuffSize = 8;
 
-WEBRTC_DEFINE_string(i, "", "Input wav file");
-WEBRTC_DEFINE_string(o, "", "VAD output file");
+DEFINE_string(i, "", "Input wav file");
+DEFINE_string(o, "", "VAD output file");
 
 int main(int argc, char* argv[]) {
   if (rtc::FlagList::SetFlagsFromCommandLine(&argc, argv, true))
diff --git a/modules/audio_processing/transient/transient_suppression_test.cc b/modules/audio_processing/transient/transient_suppression_test.cc
index e15f69c..9e7ecd5 100644
--- a/modules/audio_processing/transient/transient_suppression_test.cc
+++ b/modules/audio_processing/transient/transient_suppression_test.cc
@@ -23,28 +23,26 @@
 #include "test/gtest.h"
 #include "test/testsupport/fileutils.h"
 
-WEBRTC_DEFINE_string(in_file_name, "", "PCM file that contains the signal.");
-WEBRTC_DEFINE_string(detection_file_name,
-                     "",
-                     "PCM file that contains the detection signal.");
-WEBRTC_DEFINE_string(reference_file_name,
-                     "",
-                     "PCM file that contains the reference signal.");
+DEFINE_string(in_file_name, "", "PCM file that contains the signal.");
+DEFINE_string(detection_file_name,
+              "",
+              "PCM file that contains the detection signal.");
+DEFINE_string(reference_file_name,
+              "",
+              "PCM file that contains the reference signal.");
 
-WEBRTC_DEFINE_int(chunk_size_ms,
-                  10,
-                  "Time between each chunk of samples in milliseconds.");
+DEFINE_int(chunk_size_ms,
+           10,
+           "Time between each chunk of samples in milliseconds.");
 
-WEBRTC_DEFINE_int(sample_rate_hz,
-                  16000,
-                  "Sampling frequency of the signal in Hertz.");
-WEBRTC_DEFINE_int(detection_rate_hz,
-                  0,
-                  "Sampling frequency of the detection signal in Hertz.");
+DEFINE_int(sample_rate_hz, 16000, "Sampling frequency of the signal in Hertz.");
+DEFINE_int(detection_rate_hz,
+           0,
+           "Sampling frequency of the detection signal in Hertz.");
 
-WEBRTC_DEFINE_int(num_channels, 1, "Number of channels.");
+DEFINE_int(num_channels, 1, "Number of channels.");
 
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_bool(help, false, "Print this message.");
 
 namespace webrtc {
 
diff --git a/modules/remote_bitrate_estimator/tools/bwe_rtp.cc b/modules/remote_bitrate_estimator/tools/bwe_rtp.cc
index d7cdb54..0230db1 100644
--- a/modules/remote_bitrate_estimator/tools/bwe_rtp.cc
+++ b/modules/remote_bitrate_estimator/tools/bwe_rtp.cc
@@ -24,30 +24,28 @@
 
 namespace flags {
 
-WEBRTC_DEFINE_string(
-    extension_type,
-    "abs",
-    "Extension type, either abs for absolute send time or tsoffset "
-    "for timestamp offset.");
+DEFINE_string(extension_type,
+              "abs",
+              "Extension type, either abs for absolute send time or tsoffset "
+              "for timestamp offset.");
 std::string ExtensionType() {
   return static_cast<std::string>(FLAG_extension_type);
 }
 
-WEBRTC_DEFINE_int(extension_id, 3, "Extension id.");
+DEFINE_int(extension_id, 3, "Extension id.");
 int ExtensionId() {
   return static_cast<int>(FLAG_extension_id);
 }
 
-WEBRTC_DEFINE_string(input_file, "", "Input file.");
+DEFINE_string(input_file, "", "Input file.");
 std::string InputFile() {
   return static_cast<std::string>(FLAG_input_file);
 }
 
-WEBRTC_DEFINE_string(
-    ssrc_filter,
-    "",
-    "Comma-separated list of SSRCs in hexadecimal which are to be "
-    "used as input to the BWE (only applicable to pcap files).");
+DEFINE_string(ssrc_filter,
+              "",
+              "Comma-separated list of SSRCs in hexadecimal which are to be "
+              "used as input to the BWE (only applicable to pcap files).");
 std::set<uint32_t> SsrcFilter() {
   std::string ssrc_filter_string = static_cast<std::string>(FLAG_ssrc_filter);
   if (ssrc_filter_string.empty())
@@ -66,7 +64,7 @@
   return ssrcs;
 }
 
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_bool(help, false, "Print this message.");
 }  // namespace flags
 
 bool ParseArgsAndSetupEstimator(int argc,
diff --git a/rtc_base/flags.h b/rtc_base/flags.h
index a08bfd2..1c476d8 100644
--- a/rtc_base/flags.h
+++ b/rtc_base/flags.h
@@ -36,7 +36,7 @@
   // bool values ('bool b = "false";' results in b == true!), we pass
   // and int argument to New_BOOL as this appears to be safer - sigh.
   // In particular, it prevents the (not uncommon!) bug where a bool
-  // flag is defined via: WEBRTC_DEFINE_bool(flag, "false", "some comment");.
+  // flag is defined via: DEFINE_bool(flag, "false", "some comment");.
   static FlagValue New_BOOL(int b) {
     FlagValue v;
     v.b = (b != 0);
@@ -155,7 +155,7 @@
 };
 
 // Internal use only.
-#define WEBRTC_DEFINE_FLAG(type, c_type, name, default, comment)            \
+#define DEFINE_FLAG(type, c_type, name, default, comment)                   \
   /* define and initialize the flag */                                      \
   c_type FLAG_##name = (default);                                           \
   /* register the flag */                                                   \
@@ -164,25 +164,25 @@
                                rtc::FlagValue::New_##type(default))
 
 // Internal use only.
-#define WEBRTC_DECLARE_FLAG(c_type, name) \
-  /* declare the external flag */         \
+#define DECLARE_FLAG(c_type, name) \
+  /* declare the external flag */  \
   extern c_type FLAG_##name
 
 // Use the following macros to define a new flag:
-#define WEBRTC_DEFINE_bool(name, default, comment) \
-  WEBRTC_DEFINE_FLAG(BOOL, bool, name, default, comment)
-#define WEBRTC_DEFINE_int(name, default, comment) \
-  WEBRTC_DEFINE_FLAG(INT, int, name, default, comment)
-#define WEBRTC_DEFINE_float(name, default, comment) \
-  WEBRTC_DEFINE_FLAG(FLOAT, double, name, default, comment)
-#define WEBRTC_DEFINE_string(name, default, comment) \
-  WEBRTC_DEFINE_FLAG(STRING, const char*, name, default, comment)
+#define DEFINE_bool(name, default, comment) \
+  DEFINE_FLAG(BOOL, bool, name, default, comment)
+#define DEFINE_int(name, default, comment) \
+  DEFINE_FLAG(INT, int, name, default, comment)
+#define DEFINE_float(name, default, comment) \
+  DEFINE_FLAG(FLOAT, double, name, default, comment)
+#define DEFINE_string(name, default, comment) \
+  DEFINE_FLAG(STRING, const char*, name, default, comment)
 
 // Use the following macros to declare a flag defined elsewhere:
-#define WEBRTC_DECLARE_bool(name) WEBRTC_DECLARE_FLAG(bool, name)
-#define WEBRTC_DECLARE_int(name) WEBRTC_DECLARE_FLAG(int, name)
-#define WEBRTC_DECLARE_float(name) WEBRTC_DECLARE_FLAG(double, name)
-#define WEBRTC_DECLARE_string(name) WEBRTC_DECLARE_FLAG(const char*, name)
+#define DECLARE_bool(name) DECLARE_FLAG(bool, name)
+#define DECLARE_int(name) DECLARE_FLAG(int, name)
+#define DECLARE_float(name) DECLARE_FLAG(double, name)
+#define DECLARE_string(name) DECLARE_FLAG(const char*, name)
 
 // The global list of all flags.
 class FlagList {
diff --git a/rtc_base/unittest_main.cc b/rtc_base/unittest_main.cc
index 28a1bbb..98b8a48 100644
--- a/rtc_base/unittest_main.cc
+++ b/rtc_base/unittest_main.cc
@@ -31,20 +31,19 @@
 #include "test/ios/test_support.h"
 #endif
 
-WEBRTC_DEFINE_bool(help, false, "prints this message");
-WEBRTC_DEFINE_string(log, "", "logging options to use");
-WEBRTC_DEFINE_string(
+DEFINE_bool(help, false, "prints this message");
+DEFINE_string(log, "", "logging options to use");
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
     "E.g. running with --force_fieldtrials=WebRTC-FooFeature/Enable/"
     " will assign the group Enable to field trial WebRTC-FooFeature.");
 #if defined(WEBRTC_WIN)
-WEBRTC_DEFINE_int(crt_break_alloc, -1, "memory allocation to break on");
-WEBRTC_DEFINE_bool(
-    default_error_handlers,
-    false,
-    "leave the default exception/dbg handler functions in place");
+DEFINE_int(crt_break_alloc, -1, "memory allocation to break on");
+DEFINE_bool(default_error_handlers,
+            false,
+            "leave the default exception/dbg handler functions in place");
 
 void TestInvalidParameterHandler(const wchar_t* expression,
                                  const wchar_t* function,
diff --git a/rtc_tools/agc/activity_metric.cc b/rtc_tools/agc/activity_metric.cc
index 9b2276f..b4ed3fa 100644
--- a/rtc_tools/agc/activity_metric.cc
+++ b/rtc_tools/agc/activity_metric.cc
@@ -30,34 +30,32 @@
 static const int kAgcAnalWindowSamples = 100;
 static const float kDefaultActivityThreshold = 0.3f;
 
-WEBRTC_DEFINE_bool(standalone_vad, true, "enable stand-alone VAD");
-WEBRTC_DEFINE_string(true_vad,
-                     "",
-                     "name of a file containing true VAD in 'int'"
-                     " format");
-WEBRTC_DEFINE_string(
-    video_vad,
-    "",
-    "name of a file containing video VAD (activity"
-    " probabilities) in double format. One activity per 10ms is"
-    " required. If no file is given the video information is not"
-    " incorporated. Negative activity is interpreted as video is"
-    " not adapted and the statistics are not computed during"
-    " the learning phase. Note that the negative video activities"
-    " are ONLY allowed at the beginning.");
-WEBRTC_DEFINE_string(
-    result,
-    "",
-    "name of a file to write the results. The results"
-    " will be appended to the end of the file. This is optional.");
-WEBRTC_DEFINE_string(audio_content,
-                     "",
-                     "name of a file where audio content is written"
-                     " to, in double format.");
-WEBRTC_DEFINE_float(activity_threshold,
-                    kDefaultActivityThreshold,
-                    "Activity threshold");
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_bool(standalone_vad, true, "enable stand-alone VAD");
+DEFINE_string(true_vad,
+              "",
+              "name of a file containing true VAD in 'int'"
+              " format");
+DEFINE_string(video_vad,
+              "",
+              "name of a file containing video VAD (activity"
+              " probabilities) in double format. One activity per 10ms is"
+              " required. If no file is given the video information is not"
+              " incorporated. Negative activity is interpreted as video is"
+              " not adapted and the statistics are not computed during"
+              " the learning phase. Note that the negative video activities"
+              " are ONLY allowed at the beginning.");
+DEFINE_string(result,
+              "",
+              "name of a file to write the results. The results"
+              " will be appended to the end of the file. This is optional.");
+DEFINE_string(audio_content,
+              "",
+              "name of a file where audio content is written"
+              " to, in double format.");
+DEFINE_float(activity_threshold,
+             kDefaultActivityThreshold,
+             "Activity threshold");
+DEFINE_bool(help, false, "prints this message");
 
 namespace webrtc {
 
diff --git a/rtc_tools/event_log_visualizer/main.cc b/rtc_tools/event_log_visualizer/main.cc
index ad90cfb..2a52d06 100644
--- a/rtc_tools/event_log_visualizer/main.cc
+++ b/rtc_tools/event_log_visualizer/main.cc
@@ -20,173 +20,151 @@
 #include "test/field_trial.h"
 #include "test/testsupport/fileutils.h"
 
-WEBRTC_DEFINE_string(
-    plot_profile,
-    "default",
-    "A profile that selects a certain subset of the plots. Currently "
-    "defined profiles are \"all\", \"none\", \"sendside_bwe\","
-    "\"receiveside_bwe\" and \"default\"");
+DEFINE_string(plot_profile,
+              "default",
+              "A profile that selects a certain subset of the plots. Currently "
+              "defined profiles are \"all\", \"none\", \"sendside_bwe\","
+              "\"receiveside_bwe\" and \"default\"");
 
-WEBRTC_DEFINE_bool(plot_incoming_packet_sizes,
-                   false,
-                   "Plot bar graph showing the size of each incoming packet.");
-WEBRTC_DEFINE_bool(plot_outgoing_packet_sizes,
-                   false,
-                   "Plot bar graph showing the size of each outgoing packet.");
-WEBRTC_DEFINE_bool(
-    plot_incoming_packet_count,
-    false,
-    "Plot the accumulated number of packets for each incoming stream.");
-WEBRTC_DEFINE_bool(
-    plot_outgoing_packet_count,
-    false,
-    "Plot the accumulated number of packets for each outgoing stream.");
-WEBRTC_DEFINE_bool(
-    plot_audio_playout,
-    false,
-    "Plot bar graph showing the time between each audio playout.");
-WEBRTC_DEFINE_bool(
-    plot_audio_level,
-    false,
-    "Plot line graph showing the audio level of incoming audio.");
-WEBRTC_DEFINE_bool(
-    plot_incoming_sequence_number_delta,
-    false,
-    "Plot the sequence number difference between consecutive incoming "
-    "packets.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(plot_incoming_packet_sizes,
+            false,
+            "Plot bar graph showing the size of each incoming packet.");
+DEFINE_bool(plot_outgoing_packet_sizes,
+            false,
+            "Plot bar graph showing the size of each outgoing packet.");
+DEFINE_bool(plot_incoming_packet_count,
+            false,
+            "Plot the accumulated number of packets for each incoming stream.");
+DEFINE_bool(plot_outgoing_packet_count,
+            false,
+            "Plot the accumulated number of packets for each outgoing stream.");
+DEFINE_bool(plot_audio_playout,
+            false,
+            "Plot bar graph showing the time between each audio playout.");
+DEFINE_bool(plot_audio_level,
+            false,
+            "Plot line graph showing the audio level of incoming audio.");
+DEFINE_bool(plot_incoming_sequence_number_delta,
+            false,
+            "Plot the sequence number difference between consecutive incoming "
+            "packets.");
+DEFINE_bool(
     plot_incoming_delay_delta,
     false,
     "Plot the difference in 1-way path delay between consecutive packets.");
-WEBRTC_DEFINE_bool(
-    plot_incoming_delay,
-    true,
-    "Plot the 1-way path delay for incoming packets, normalized so "
-    "that the first packet has delay 0.");
-WEBRTC_DEFINE_bool(
-    plot_incoming_loss_rate,
-    true,
-    "Compute the loss rate for incoming packets using a method that's "
-    "similar to the one used for RTCP SR and RR fraction lost. Note "
-    "that the loss rate can be negative if packets are duplicated or "
-    "reordered.");
-WEBRTC_DEFINE_bool(plot_incoming_bitrate,
-                   true,
-                   "Plot the total bitrate used by all incoming streams.");
-WEBRTC_DEFINE_bool(plot_outgoing_bitrate,
-                   true,
-                   "Plot the total bitrate used by all outgoing streams.");
-WEBRTC_DEFINE_bool(plot_incoming_stream_bitrate,
-                   true,
-                   "Plot the bitrate used by each incoming stream.");
-WEBRTC_DEFINE_bool(plot_outgoing_stream_bitrate,
-                   true,
-                   "Plot the bitrate used by each outgoing stream.");
-WEBRTC_DEFINE_bool(
-    plot_simulated_receiveside_bwe,
-    false,
-    "Run the receive-side bandwidth estimator with the incoming rtp "
-    "packets and plot the resulting estimate.");
-WEBRTC_DEFINE_bool(
-    plot_simulated_sendside_bwe,
-    false,
-    "Run the send-side bandwidth estimator with the outgoing rtp and "
-    "incoming rtcp and plot the resulting estimate.");
-WEBRTC_DEFINE_bool(
-    plot_network_delay_feedback,
-    true,
-    "Compute network delay based on sent packets and the received "
-    "transport feedback.");
-WEBRTC_DEFINE_bool(
-    plot_fraction_loss_feedback,
-    true,
-    "Plot packet loss in percent for outgoing packets (as perceived by "
-    "the send-side bandwidth estimator).");
-WEBRTC_DEFINE_bool(
-    plot_pacer_delay,
-    false,
-    "Plot the time each sent packet has spent in the pacer (based on "
-    "the difference between the RTP timestamp and the send "
-    "timestamp).");
-WEBRTC_DEFINE_bool(
-    plot_timestamps,
-    false,
-    "Plot the rtp timestamps of all rtp and rtcp packets over time.");
-WEBRTC_DEFINE_bool(
-    plot_rtcp_details,
-    false,
-    "Plot the contents of all report blocks in all sender and receiver "
-    "reports. This includes fraction lost, cumulative number of lost "
-    "packets, extended highest sequence number and time since last "
-    "received SR.");
-WEBRTC_DEFINE_bool(plot_audio_encoder_bitrate_bps,
-                   false,
-                   "Plot the audio encoder target bitrate.");
-WEBRTC_DEFINE_bool(plot_audio_encoder_frame_length_ms,
-                   false,
-                   "Plot the audio encoder frame length.");
-WEBRTC_DEFINE_bool(
+DEFINE_bool(plot_incoming_delay,
+            true,
+            "Plot the 1-way path delay for incoming packets, normalized so "
+            "that the first packet has delay 0.");
+DEFINE_bool(plot_incoming_loss_rate,
+            true,
+            "Compute the loss rate for incoming packets using a method that's "
+            "similar to the one used for RTCP SR and RR fraction lost. Note "
+            "that the loss rate can be negative if packets are duplicated or "
+            "reordered.");
+DEFINE_bool(plot_incoming_bitrate,
+            true,
+            "Plot the total bitrate used by all incoming streams.");
+DEFINE_bool(plot_outgoing_bitrate,
+            true,
+            "Plot the total bitrate used by all outgoing streams.");
+DEFINE_bool(plot_incoming_stream_bitrate,
+            true,
+            "Plot the bitrate used by each incoming stream.");
+DEFINE_bool(plot_outgoing_stream_bitrate,
+            true,
+            "Plot the bitrate used by each outgoing stream.");
+DEFINE_bool(plot_simulated_receiveside_bwe,
+            false,
+            "Run the receive-side bandwidth estimator with the incoming rtp "
+            "packets and plot the resulting estimate.");
+DEFINE_bool(plot_simulated_sendside_bwe,
+            false,
+            "Run the send-side bandwidth estimator with the outgoing rtp and "
+            "incoming rtcp and plot the resulting estimate.");
+DEFINE_bool(plot_network_delay_feedback,
+            true,
+            "Compute network delay based on sent packets and the received "
+            "transport feedback.");
+DEFINE_bool(plot_fraction_loss_feedback,
+            true,
+            "Plot packet loss in percent for outgoing packets (as perceived by "
+            "the send-side bandwidth estimator).");
+DEFINE_bool(plot_pacer_delay,
+            false,
+            "Plot the time each sent packet has spent in the pacer (based on "
+            "the difference between the RTP timestamp and the send "
+            "timestamp).");
+DEFINE_bool(plot_timestamps,
+            false,
+            "Plot the rtp timestamps of all rtp and rtcp packets over time.");
+DEFINE_bool(plot_rtcp_details,
+            false,
+            "Plot the contents of all report blocks in all sender and receiver "
+            "reports. This includes fraction lost, cumulative number of lost "
+            "packets, extended highest sequence number and time since last "
+            "received SR.");
+DEFINE_bool(plot_audio_encoder_bitrate_bps,
+            false,
+            "Plot the audio encoder target bitrate.");
+DEFINE_bool(plot_audio_encoder_frame_length_ms,
+            false,
+            "Plot the audio encoder frame length.");
+DEFINE_bool(
     plot_audio_encoder_packet_loss,
     false,
     "Plot the uplink packet loss fraction which is sent to the audio encoder.");
-WEBRTC_DEFINE_bool(plot_audio_encoder_fec,
-                   false,
-                   "Plot the audio encoder FEC.");
-WEBRTC_DEFINE_bool(plot_audio_encoder_dtx,
-                   false,
-                   "Plot the audio encoder DTX.");
-WEBRTC_DEFINE_bool(plot_audio_encoder_num_channels,
-                   false,
-                   "Plot the audio encoder number of channels.");
-WEBRTC_DEFINE_bool(plot_neteq_stats, false, "Plot the NetEq statistics.");
-WEBRTC_DEFINE_bool(plot_ice_candidate_pair_config,
-                   false,
-                   "Plot the ICE candidate pair config events.");
-WEBRTC_DEFINE_bool(plot_ice_connectivity_check,
-                   false,
-                   "Plot the ICE candidate pair connectivity checks.");
+DEFINE_bool(plot_audio_encoder_fec, false, "Plot the audio encoder FEC.");
+DEFINE_bool(plot_audio_encoder_dtx, false, "Plot the audio encoder DTX.");
+DEFINE_bool(plot_audio_encoder_num_channels,
+            false,
+            "Plot the audio encoder number of channels.");
+DEFINE_bool(plot_neteq_stats, false, "Plot the NetEq statistics.");
+DEFINE_bool(plot_ice_candidate_pair_config,
+            false,
+            "Plot the ICE candidate pair config events.");
+DEFINE_bool(plot_ice_connectivity_check,
+            false,
+            "Plot the ICE candidate pair connectivity checks.");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
     "E.g. running with --force_fieldtrials=WebRTC-FooFeature/Enabled/"
     " will assign the group Enabled to field trial WebRTC-FooFeature. Multiple "
     "trials are separated by \"/\"");
-WEBRTC_DEFINE_string(wav_filename,
-                     "",
-                     "Path to wav file used for simulation of jitter buffer");
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_string(wav_filename,
+              "",
+              "Path to wav file used for simulation of jitter buffer");
+DEFINE_bool(help, false, "prints this message");
 
-WEBRTC_DEFINE_bool(
-    show_detector_state,
-    false,
-    "Show the state of the delay based BWE detector on the total "
-    "bitrate graph");
+DEFINE_bool(show_detector_state,
+            false,
+            "Show the state of the delay based BWE detector on the total "
+            "bitrate graph");
 
-WEBRTC_DEFINE_bool(show_alr_state,
-                   false,
-                   "Show the state ALR state on the total bitrate graph");
+DEFINE_bool(show_alr_state,
+            false,
+            "Show the state ALR state on the total bitrate graph");
 
-WEBRTC_DEFINE_bool(
-    parse_unconfigured_header_extensions,
-    true,
-    "Attempt to parse unconfigured header extensions using the default "
-    "WebRTC mapping. This can give very misleading results if the "
-    "application negotiates a different mapping.");
+DEFINE_bool(parse_unconfigured_header_extensions,
+            true,
+            "Attempt to parse unconfigured header extensions using the default "
+            "WebRTC mapping. This can give very misleading results if the "
+            "application negotiates a different mapping.");
 
-WEBRTC_DEFINE_bool(print_triage_alerts,
-                   false,
-                   "Print triage alerts, i.e. a list of potential problems.");
+DEFINE_bool(print_triage_alerts,
+            false,
+            "Print triage alerts, i.e. a list of potential problems.");
 
-WEBRTC_DEFINE_bool(
-    normalize_time,
-    true,
-    "Normalize the log timestamps so that the call starts at time 0.");
+DEFINE_bool(normalize_time,
+            true,
+            "Normalize the log timestamps so that the call starts at time 0.");
 
-WEBRTC_DEFINE_bool(protobuf_output,
-                   false,
-                   "Output charts as protobuf instead of python code.");
+DEFINE_bool(protobuf_output,
+            false,
+            "Output charts as protobuf instead of python code.");
 
 void SetAllPlotFlags(bool setting);
 
diff --git a/rtc_tools/unpack_aecdump/unpack.cc b/rtc_tools/unpack_aecdump/unpack.cc
index 142b497..af8bf31 100644
--- a/rtc_tools/unpack_aecdump/unpack.cc
+++ b/rtc_tools/unpack_aecdump/unpack.cc
@@ -29,34 +29,27 @@
 RTC_POP_IGNORING_WUNDEF()
 
 // TODO(andrew): unpack more of the data.
-WEBRTC_DEFINE_string(input_file, "input", "The name of the input stream file.");
-WEBRTC_DEFINE_string(output_file,
-                     "ref_out",
-                     "The name of the reference output stream file.");
-WEBRTC_DEFINE_string(reverse_file,
-                     "reverse",
-                     "The name of the reverse input stream file.");
-WEBRTC_DEFINE_string(delay_file, "delay.int32", "The name of the delay file.");
-WEBRTC_DEFINE_string(drift_file, "drift.int32", "The name of the drift file.");
-WEBRTC_DEFINE_string(level_file, "level.int32", "The name of the level file.");
-WEBRTC_DEFINE_string(keypress_file,
-                     "keypress.bool",
-                     "The name of the keypress file.");
-WEBRTC_DEFINE_string(callorder_file,
-                     "callorder",
-                     "The name of the render/capture call order file.");
-WEBRTC_DEFINE_string(settings_file,
-                     "settings.txt",
-                     "The name of the settings file.");
-WEBRTC_DEFINE_bool(full,
-                   false,
-                   "Unpack the full set of files (normally not needed).");
-WEBRTC_DEFINE_bool(raw, false, "Write raw data instead of a WAV file.");
-WEBRTC_DEFINE_bool(
-    text,
-    false,
-    "Write non-audio files as text files instead of binary files.");
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_string(input_file, "input", "The name of the input stream file.");
+DEFINE_string(output_file,
+              "ref_out",
+              "The name of the reference output stream file.");
+DEFINE_string(reverse_file,
+              "reverse",
+              "The name of the reverse input stream file.");
+DEFINE_string(delay_file, "delay.int32", "The name of the delay file.");
+DEFINE_string(drift_file, "drift.int32", "The name of the drift file.");
+DEFINE_string(level_file, "level.int32", "The name of the level file.");
+DEFINE_string(keypress_file, "keypress.bool", "The name of the keypress file.");
+DEFINE_string(callorder_file,
+              "callorder",
+              "The name of the render/capture call order file.");
+DEFINE_string(settings_file, "settings.txt", "The name of the settings file.");
+DEFINE_bool(full, false, "Unpack the full set of files (normally not needed).");
+DEFINE_bool(raw, false, "Write raw data instead of a WAV file.");
+DEFINE_bool(text,
+            false,
+            "Write non-audio files as text files instead of binary files.");
+DEFINE_bool(help, false, "Print this message.");
 
 #define PRINT_CONFIG(field_name)                                         \
   if (msg.has_##field_name()) {                                          \
diff --git a/test/scenario/scenario.cc b/test/scenario/scenario.cc
index 2395a7a..7eb09e8 100644
--- a/test/scenario/scenario.cc
+++ b/test/scenario/scenario.cc
@@ -16,7 +16,7 @@
 #include "rtc_base/flags.h"
 #include "test/testsupport/fileutils.h"
 
-WEBRTC_DEFINE_bool(scenario_logs, false, "Save logs from scenario framework.");
+DEFINE_bool(scenario_logs, false, "Save logs from scenario framework.");
 
 namespace webrtc {
 namespace test {
diff --git a/test/test_main_lib.cc b/test/test_main_lib.cc
index 95144c9..8e1b213 100644
--- a/test/test_main_lib.cc
+++ b/test/test_main_lib.cc
@@ -32,13 +32,13 @@
 #if defined(WEBRTC_IOS)
 #include "test/ios/test_support.h"
 
-WEBRTC_DEFINE_string(NSTreatUnknownArgumentsAsOpen,
-                     "",
-                     "Intentionally ignored flag intended for iOS simulator.");
-WEBRTC_DEFINE_string(ApplePersistenceIgnoreState,
-                     "",
-                     "Intentionally ignored flag intended for iOS simulator.");
-WEBRTC_DEFINE_bool(
+DEFINE_string(NSTreatUnknownArgumentsAsOpen,
+              "",
+              "Intentionally ignored flag intended for iOS simulator.");
+DEFINE_string(ApplePersistenceIgnoreState,
+              "",
+              "Intentionally ignored flag intended for iOS simulator.");
+DEFINE_bool(
     save_chartjson_result,
     false,
     "Store the perf results in Documents/perf_result.json in the format "
@@ -48,12 +48,12 @@
 
 #else
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     isolated_script_test_output,
     "",
     "Path to output an empty JSON file which Chromium infra requires.");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     isolated_script_test_perf_output,
     "",
     "Path where the perf results should be stored in the JSON format described "
@@ -63,16 +63,16 @@
 
 #endif
 
-WEBRTC_DEFINE_bool(logs, false, "print logs to stderr");
+DEFINE_bool(logs, false, "print logs to stderr");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
     "E.g. running with --force_fieldtrials=WebRTC-FooFeature/Enable/"
     " will assign the group Enable to field trial WebRTC-FooFeature.");
 
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_bool(help, false, "Print this message.");
 
 namespace webrtc {
 
diff --git a/test/testsupport/test_artifacts.cc b/test/testsupport/test_artifacts.cc
index 9438cef..0f7e0cd 100644
--- a/test/testsupport/test_artifacts.cc
+++ b/test/testsupport/test_artifacts.cc
@@ -24,9 +24,9 @@
 }
 }  // namespace
 
-WEBRTC_DEFINE_string(test_artifacts_dir,
-                     DefaultArtifactPath().c_str(),
-                     "The output folder where test output should be saved.");
+DEFINE_string(test_artifacts_dir,
+              DefaultArtifactPath().c_str(),
+              "The output folder where test output should be saved.");
 
 namespace webrtc {
 namespace test {
diff --git a/test/testsupport/test_artifacts_unittest.cc b/test/testsupport/test_artifacts_unittest.cc
index 267ea93..c423cd9 100644
--- a/test/testsupport/test_artifacts_unittest.cc
+++ b/test/testsupport/test_artifacts_unittest.cc
@@ -21,7 +21,7 @@
 #include "test/gtest.h"
 #include "test/testsupport/fileutils.h"
 
-WEBRTC_DECLARE_string(test_artifacts_dir);
+DECLARE_string(test_artifacts_dir);
 
 namespace webrtc {
 namespace test {
diff --git a/video/full_stack_tests.cc b/video/full_stack_tests.cc
index 6562173..21a8e62 100644
--- a/video/full_stack_tests.cc
+++ b/video/full_stack_tests.cc
@@ -22,24 +22,21 @@
 namespace webrtc {
 namespace flags {
 
-WEBRTC_DEFINE_string(rtc_event_log_name,
-                     "",
-                     "Filename for rtc event log. Two files "
-                     "with \"_send\" and \"_recv\" suffixes will be created.");
+DEFINE_string(rtc_event_log_name,
+              "",
+              "Filename for rtc event log. Two files "
+              "with \"_send\" and \"_recv\" suffixes will be created.");
 std::string RtcEventLogName() {
   return static_cast<std::string>(FLAG_rtc_event_log_name);
 }
-WEBRTC_DEFINE_string(rtp_dump_name,
-                     "",
-                     "Filename for dumped received RTP stream.");
+DEFINE_string(rtp_dump_name, "", "Filename for dumped received RTP stream.");
 std::string RtpDumpName() {
   return static_cast<std::string>(FLAG_rtp_dump_name);
 }
-WEBRTC_DEFINE_string(
-    encoded_frame_path,
-    "",
-    "The base path for encoded frame logs. Created files will have "
-    "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
+DEFINE_string(encoded_frame_path,
+              "",
+              "The base path for encoded frame logs. Created files will have "
+              "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
 std::string EncodedFramePath() {
   return static_cast<std::string>(FLAG_encoded_frame_path);
 }
diff --git a/video/replay.cc b/video/replay.cc
index a3fd175..5c17e5b 100644
--- a/video/replay.cc
+++ b/video/replay.cc
@@ -72,39 +72,39 @@
 // TODO(pbos): Multiple receivers.
 
 // Flag for payload type.
-WEBRTC_DEFINE_int(media_payload_type,
-                  test::CallTest::kPayloadTypeVP8,
-                  "Media payload type");
+DEFINE_int(media_payload_type,
+           test::CallTest::kPayloadTypeVP8,
+           "Media payload type");
 static int MediaPayloadType() {
   return static_cast<int>(FLAG_media_payload_type);
 }
 
 // Flag for RED payload type.
-WEBRTC_DEFINE_int(red_payload_type,
-                  test::CallTest::kRedPayloadType,
-                  "RED payload type");
+DEFINE_int(red_payload_type,
+           test::CallTest::kRedPayloadType,
+           "RED payload type");
 static int RedPayloadType() {
   return static_cast<int>(FLAG_red_payload_type);
 }
 
 // Flag for ULPFEC payload type.
-WEBRTC_DEFINE_int(ulpfec_payload_type,
-                  test::CallTest::kUlpfecPayloadType,
-                  "ULPFEC payload type");
+DEFINE_int(ulpfec_payload_type,
+           test::CallTest::kUlpfecPayloadType,
+           "ULPFEC payload type");
 static int UlpfecPayloadType() {
   return static_cast<int>(FLAG_ulpfec_payload_type);
 }
 
-WEBRTC_DEFINE_int(media_payload_type_rtx,
-                  test::CallTest::kSendRtxPayloadType,
-                  "Media over RTX payload type");
+DEFINE_int(media_payload_type_rtx,
+           test::CallTest::kSendRtxPayloadType,
+           "Media over RTX payload type");
 static int MediaPayloadTypeRtx() {
   return static_cast<int>(FLAG_media_payload_type_rtx);
 }
 
-WEBRTC_DEFINE_int(red_payload_type_rtx,
-                  test::CallTest::kRtxRedPayloadType,
-                  "RED over RTX payload type");
+DEFINE_int(red_payload_type_rtx,
+           test::CallTest::kRtxRedPayloadType,
+           "RED over RTX payload type");
 static int RedPayloadTypeRtx() {
   return static_cast<int>(FLAG_red_payload_type_rtx);
 }
@@ -115,7 +115,7 @@
       std::to_string(test::CallTest::kVideoSendSsrcs[0]);
   return ssrc;
 }
-WEBRTC_DEFINE_string(ssrc, DefaultSsrc().c_str(), "Incoming SSRC");
+DEFINE_string(ssrc, DefaultSsrc().c_str(), "Incoming SSRC");
 static uint32_t Ssrc() {
   return rtc::StringToNumber<uint32_t>(FLAG_ssrc).value();
 }
@@ -125,56 +125,54 @@
       std::to_string(test::CallTest::kSendRtxSsrcs[0]);
   return ssrc_rtx;
 }
-WEBRTC_DEFINE_string(ssrc_rtx, DefaultSsrcRtx().c_str(), "Incoming RTX SSRC");
+DEFINE_string(ssrc_rtx, DefaultSsrcRtx().c_str(), "Incoming RTX SSRC");
 static uint32_t SsrcRtx() {
   return rtc::StringToNumber<uint32_t>(FLAG_ssrc_rtx).value();
 }
 
 // Flag for abs-send-time id.
-WEBRTC_DEFINE_int(abs_send_time_id, -1, "RTP extension ID for abs-send-time");
+DEFINE_int(abs_send_time_id, -1, "RTP extension ID for abs-send-time");
 static int AbsSendTimeId() {
   return static_cast<int>(FLAG_abs_send_time_id);
 }
 
 // Flag for transmission-offset id.
-WEBRTC_DEFINE_int(transmission_offset_id,
-                  -1,
-                  "RTP extension ID for transmission-offset");
+DEFINE_int(transmission_offset_id,
+           -1,
+           "RTP extension ID for transmission-offset");
 static int TransmissionOffsetId() {
   return static_cast<int>(FLAG_transmission_offset_id);
 }
 
 // Flag for rtpdump input file.
-WEBRTC_DEFINE_string(input_file, "", "input file");
+DEFINE_string(input_file, "", "input file");
 static std::string InputFile() {
   return static_cast<std::string>(FLAG_input_file);
 }
 
-WEBRTC_DEFINE_string(config_file, "", "config file");
+DEFINE_string(config_file, "", "config file");
 static std::string ConfigFile() {
   return static_cast<std::string>(FLAG_config_file);
 }
 
 // Flag for raw output files.
-WEBRTC_DEFINE_string(out_base, "", "Basename (excluding .jpg) for raw output");
+DEFINE_string(out_base, "", "Basename (excluding .jpg) for raw output");
 static std::string OutBase() {
   return static_cast<std::string>(FLAG_out_base);
 }
 
-WEBRTC_DEFINE_string(decoder_bitstream_filename,
-                     "",
-                     "Decoder bitstream output file");
+DEFINE_string(decoder_bitstream_filename, "", "Decoder bitstream output file");
 static std::string DecoderBitstreamFilename() {
   return static_cast<std::string>(FLAG_decoder_bitstream_filename);
 }
 
 // Flag for video codec.
-WEBRTC_DEFINE_string(codec, "VP8", "Video codec");
+DEFINE_string(codec, "VP8", "Video codec");
 static std::string Codec() {
   return static_cast<std::string>(FLAG_codec);
 }
 
-WEBRTC_DEFINE_bool(help, false, "Print this message.");
+DEFINE_bool(help, false, "Print this message.");
 }  // namespace flags
 
 static const uint32_t kReceiverLocalSsrc = 0x123456;
diff --git a/video/screenshare_loopback.cc b/video/screenshare_loopback.cc
index 4696666..245d263 100644
--- a/video/screenshare_loopback.cc
+++ b/video/screenshare_loopback.cc
@@ -22,76 +22,73 @@
 namespace flags {
 
 // Flags common with video loopback, with different default values.
-WEBRTC_DEFINE_int(width, 1850, "Video width (crops source).");
+DEFINE_int(width, 1850, "Video width (crops source).");
 size_t Width() {
   return static_cast<size_t>(FLAG_width);
 }
 
-WEBRTC_DEFINE_int(height, 1110, "Video height (crops source).");
+DEFINE_int(height, 1110, "Video height (crops source).");
 size_t Height() {
   return static_cast<size_t>(FLAG_height);
 }
 
-WEBRTC_DEFINE_int(fps, 5, "Frames per second.");
+DEFINE_int(fps, 5, "Frames per second.");
 int Fps() {
   return static_cast<int>(FLAG_fps);
 }
 
-WEBRTC_DEFINE_int(min_bitrate, 50, "Call and stream min bitrate in kbps.");
+DEFINE_int(min_bitrate, 50, "Call and stream min bitrate in kbps.");
 int MinBitrateKbps() {
   return static_cast<int>(FLAG_min_bitrate);
 }
 
-WEBRTC_DEFINE_int(start_bitrate, 300, "Call start bitrate in kbps.");
+DEFINE_int(start_bitrate, 300, "Call start bitrate in kbps.");
 int StartBitrateKbps() {
   return static_cast<int>(FLAG_start_bitrate);
 }
 
-WEBRTC_DEFINE_int(target_bitrate, 200, "Stream target bitrate in kbps.");
+DEFINE_int(target_bitrate, 200, "Stream target bitrate in kbps.");
 int TargetBitrateKbps() {
   return static_cast<int>(FLAG_target_bitrate);
 }
 
-WEBRTC_DEFINE_int(max_bitrate, 1000, "Call and stream max bitrate in kbps.");
+DEFINE_int(max_bitrate, 1000, "Call and stream max bitrate in kbps.");
 int MaxBitrateKbps() {
   return static_cast<int>(FLAG_max_bitrate);
 }
 
-WEBRTC_DEFINE_int(num_temporal_layers, 2, "Number of temporal layers to use.");
+DEFINE_int(num_temporal_layers, 2, "Number of temporal layers to use.");
 int NumTemporalLayers() {
   return static_cast<int>(FLAG_num_temporal_layers);
 }
 
 // Flags common with video loopback, with equal default values.
-WEBRTC_DEFINE_string(codec, "VP8", "Video codec to use.");
+DEFINE_string(codec, "VP8", "Video codec to use.");
 std::string Codec() {
   return static_cast<std::string>(FLAG_codec);
 }
 
-WEBRTC_DEFINE_string(rtc_event_log_name,
-                     "",
-                     "Filename for rtc event log. Two files "
-                     "with \"_send\" and \"_recv\" suffixes will be created.");
+DEFINE_string(rtc_event_log_name,
+              "",
+              "Filename for rtc event log. Two files "
+              "with \"_send\" and \"_recv\" suffixes will be created.");
 std::string RtcEventLogName() {
   return static_cast<std::string>(FLAG_rtc_event_log_name);
 }
 
-WEBRTC_DEFINE_string(rtp_dump_name,
-                     "",
-                     "Filename for dumped received RTP stream.");
+DEFINE_string(rtp_dump_name, "", "Filename for dumped received RTP stream.");
 std::string RtpDumpName() {
   return static_cast<std::string>(FLAG_rtp_dump_name);
 }
 
-WEBRTC_DEFINE_int(
-    selected_tl,
-    -1,
-    "Temporal layer to show or analyze. -1 to disable filtering.");
+DEFINE_int(selected_tl,
+           -1,
+           "Temporal layer to show or analyze. -1 to disable filtering.");
 int SelectedTL() {
   return static_cast<int>(FLAG_selected_tl);
 }
 
-WEBRTC_DEFINE_int(
+DEFINE_int(
     duration,
     0,
     "Duration of the test in seconds. If 0, rendered will be shown instead.");
@@ -99,74 +96,71 @@
   return static_cast<int>(FLAG_duration);
 }
 
-WEBRTC_DEFINE_string(output_filename, "", "Target graph data filename.");
+DEFINE_string(output_filename, "", "Target graph data filename.");
 std::string OutputFilename() {
   return static_cast<std::string>(FLAG_output_filename);
 }
 
-WEBRTC_DEFINE_string(graph_title,
-                     "",
-                     "If empty, title will be generated automatically.");
+DEFINE_string(graph_title,
+              "",
+              "If empty, title will be generated automatically.");
 std::string GraphTitle() {
   return static_cast<std::string>(FLAG_graph_title);
 }
 
-WEBRTC_DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
+DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
 int LossPercent() {
   return static_cast<int>(FLAG_loss_percent);
 }
 
-WEBRTC_DEFINE_int(link_capacity,
-                  0,
-                  "Capacity (kbps) of the fake link. 0 means infinite.");
+DEFINE_int(link_capacity,
+           0,
+           "Capacity (kbps) of the fake link. 0 means infinite.");
 int LinkCapacityKbps() {
   return static_cast<int>(FLAG_link_capacity);
 }
 
-WEBRTC_DEFINE_int(queue_size,
-                  0,
-                  "Size of the bottleneck link queue in packets.");
+DEFINE_int(queue_size, 0, "Size of the bottleneck link queue in packets.");
 int QueueSize() {
   return static_cast<int>(FLAG_queue_size);
 }
 
-WEBRTC_DEFINE_int(avg_propagation_delay_ms,
-                  0,
-                  "Average link propagation delay in ms.");
+DEFINE_int(avg_propagation_delay_ms,
+           0,
+           "Average link propagation delay in ms.");
 int AvgPropagationDelayMs() {
   return static_cast<int>(FLAG_avg_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_int(std_propagation_delay_ms,
-                  0,
-                  "Link propagation delay standard deviation in ms.");
+DEFINE_int(std_propagation_delay_ms,
+           0,
+           "Link propagation delay standard deviation in ms.");
 int StdPropagationDelayMs() {
   return static_cast<int>(FLAG_std_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_int(num_streams, 0, "Number of streams to show or analyze.");
+DEFINE_int(num_streams, 0, "Number of streams to show or analyze.");
 int NumStreams() {
   return static_cast<int>(FLAG_num_streams);
 }
 
-WEBRTC_DEFINE_int(selected_stream,
-                  0,
-                  "ID of the stream to show or analyze. "
-                  "Set to the number of streams to show them all.");
+DEFINE_int(selected_stream,
+           0,
+           "ID of the stream to show or analyze. "
+           "Set to the number of streams to show them all.");
 int SelectedStream() {
   return static_cast<int>(FLAG_selected_stream);
 }
 
-WEBRTC_DEFINE_int(num_spatial_layers, 1, "Number of spatial layers to use.");
+DEFINE_int(num_spatial_layers, 1, "Number of spatial layers to use.");
 int NumSpatialLayers() {
   return static_cast<int>(FLAG_num_spatial_layers);
 }
 
-WEBRTC_DEFINE_int(
-    inter_layer_pred,
-    0,
-    "Inter-layer prediction mode. "
-    "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
+DEFINE_int(inter_layer_pred,
+           0,
+           "Inter-layer prediction mode. "
+           "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
 InterLayerPredMode InterLayerPred() {
   if (FLAG_inter_layer_pred == 0) {
     return InterLayerPredMode::kOn;
@@ -178,65 +172,58 @@
   }
 }
 
-WEBRTC_DEFINE_int(selected_sl,
-                  -1,
-                  "Spatial layer to show or analyze. -1 to disable filtering.");
+DEFINE_int(selected_sl,
+           -1,
+           "Spatial layer to show or analyze. -1 to disable filtering.");
 int SelectedSL() {
   return static_cast<int>(FLAG_selected_sl);
 }
 
-WEBRTC_DEFINE_string(
-    stream0,
-    "",
-    "Comma separated values describing VideoStream for stream #0.");
+DEFINE_string(stream0,
+              "",
+              "Comma separated values describing VideoStream for stream #0.");
 std::string Stream0() {
   return static_cast<std::string>(FLAG_stream0);
 }
 
-WEBRTC_DEFINE_string(
-    stream1,
-    "",
-    "Comma separated values describing VideoStream for stream #1.");
+DEFINE_string(stream1,
+              "",
+              "Comma separated values describing VideoStream for stream #1.");
 std::string Stream1() {
   return static_cast<std::string>(FLAG_stream1);
 }
 
-WEBRTC_DEFINE_string(
-    sl0,
-    "",
-    "Comma separated values describing SpatialLayer for layer #0.");
+DEFINE_string(sl0,
+              "",
+              "Comma separated values describing SpatialLayer for layer #0.");
 std::string SL0() {
   return static_cast<std::string>(FLAG_sl0);
 }
 
-WEBRTC_DEFINE_string(
-    sl1,
-    "",
-    "Comma separated values describing SpatialLayer for layer #1.");
+DEFINE_string(sl1,
+              "",
+              "Comma separated values describing SpatialLayer for layer #1.");
 std::string SL1() {
   return static_cast<std::string>(FLAG_sl1);
 }
 
-WEBRTC_DEFINE_string(
-    encoded_frame_path,
-    "",
-    "The base path for encoded frame logs. Created files will have "
-    "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
+DEFINE_string(encoded_frame_path,
+              "",
+              "The base path for encoded frame logs. Created files will have "
+              "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
 std::string EncodedFramePath() {
   return static_cast<std::string>(FLAG_encoded_frame_path);
 }
 
-WEBRTC_DEFINE_bool(logs, false, "print logs to stderr");
+DEFINE_bool(logs, false, "print logs to stderr");
 
-WEBRTC_DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
+DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
 
-WEBRTC_DEFINE_bool(generic_descriptor,
-                   false,
-                   "Use the generic frame descriptor.");
+DEFINE_bool(generic_descriptor, false, "Use the generic frame descriptor.");
 
-WEBRTC_DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
+DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
@@ -245,14 +232,12 @@
     "trials are separated by \"/\"");
 
 // Screenshare-specific flags.
-WEBRTC_DEFINE_int(min_transmit_bitrate,
-                  400,
-                  "Min transmit bitrate incl. padding.");
+DEFINE_int(min_transmit_bitrate, 400, "Min transmit bitrate incl. padding.");
 int MinTransmitBitrateKbps() {
   return FLAG_min_transmit_bitrate;
 }
 
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     generate_slides,
     false,
     "Whether to use randomly generated slides or read them from files.");
@@ -260,14 +245,14 @@
   return static_cast<int>(FLAG_generate_slides);
 }
 
-WEBRTC_DEFINE_int(slide_change_interval,
-                  10,
-                  "Interval (in seconds) between simulated slide changes.");
+DEFINE_int(slide_change_interval,
+           10,
+           "Interval (in seconds) between simulated slide changes.");
 int SlideChangeInterval() {
   return static_cast<int>(FLAG_slide_change_interval);
 }
 
-WEBRTC_DEFINE_int(
+DEFINE_int(
     scroll_duration,
     0,
     "Duration (in seconds) during which a slide will be scrolled into place.");
@@ -275,10 +260,9 @@
   return static_cast<int>(FLAG_scroll_duration);
 }
 
-WEBRTC_DEFINE_string(
-    slides,
-    "",
-    "Comma-separated list of *.yuv files to display as slides.");
+DEFINE_string(slides,
+              "",
+              "Comma-separated list of *.yuv files to display as slides.");
 std::vector<std::string> Slides() {
   std::vector<std::string> slides;
   std::string slides_list = FLAG_slides;
@@ -286,7 +270,7 @@
   return slides;
 }
 
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_bool(help, false, "prints this message");
 
 }  // namespace flags
 
diff --git a/video/sv_loopback.cc b/video/sv_loopback.cc
index 1ff4956..864fc63 100644
--- a/video/sv_loopback.cc
+++ b/video/sv_loopback.cc
@@ -33,79 +33,73 @@
 }
 
 // Flags for video.
-WEBRTC_DEFINE_int(vwidth, 640, "Video width.");
+DEFINE_int(vwidth, 640, "Video width.");
 size_t VideoWidth() {
   return static_cast<size_t>(FLAG_vwidth);
 }
 
-WEBRTC_DEFINE_int(vheight, 480, "Video height.");
+DEFINE_int(vheight, 480, "Video height.");
 size_t VideoHeight() {
   return static_cast<size_t>(FLAG_vheight);
 }
 
-WEBRTC_DEFINE_int(vfps, 30, "Video frames per second.");
+DEFINE_int(vfps, 30, "Video frames per second.");
 int VideoFps() {
   return static_cast<int>(FLAG_vfps);
 }
 
-WEBRTC_DEFINE_int(capture_device_index,
-                  0,
-                  "Capture device to select for video stream");
+DEFINE_int(capture_device_index,
+           0,
+           "Capture device to select for video stream");
 size_t GetCaptureDevice() {
   return static_cast<size_t>(FLAG_capture_device_index);
 }
 
-WEBRTC_DEFINE_int(vtarget_bitrate, 400, "Video stream target bitrate in kbps.");
+DEFINE_int(vtarget_bitrate, 400, "Video stream target bitrate in kbps.");
 int VideoTargetBitrateKbps() {
   return static_cast<int>(FLAG_vtarget_bitrate);
 }
 
-WEBRTC_DEFINE_int(vmin_bitrate, 100, "Video stream min bitrate in kbps.");
+DEFINE_int(vmin_bitrate, 100, "Video stream min bitrate in kbps.");
 int VideoMinBitrateKbps() {
   return static_cast<int>(FLAG_vmin_bitrate);
 }
 
-WEBRTC_DEFINE_int(vmax_bitrate, 2000, "Video stream max bitrate in kbps.");
+DEFINE_int(vmax_bitrate, 2000, "Video stream max bitrate in kbps.");
 int VideoMaxBitrateKbps() {
   return static_cast<int>(FLAG_vmax_bitrate);
 }
 
-WEBRTC_DEFINE_bool(suspend_below_min_bitrate,
-                   false,
-                   "Suspends video below the configured min bitrate.");
+DEFINE_bool(suspend_below_min_bitrate,
+            false,
+            "Suspends video below the configured min bitrate.");
 
-WEBRTC_DEFINE_int(
-    vnum_temporal_layers,
-    1,
-    "Number of temporal layers for video. Set to 1-4 to override.");
+DEFINE_int(vnum_temporal_layers,
+           1,
+           "Number of temporal layers for video. Set to 1-4 to override.");
 int VideoNumTemporalLayers() {
   return static_cast<int>(FLAG_vnum_temporal_layers);
 }
 
-WEBRTC_DEFINE_int(vnum_streams,
-                  0,
-                  "Number of video streams to show or analyze.");
+DEFINE_int(vnum_streams, 0, "Number of video streams to show or analyze.");
 int VideoNumStreams() {
   return static_cast<int>(FLAG_vnum_streams);
 }
 
-WEBRTC_DEFINE_int(vnum_spatial_layers,
-                  1,
-                  "Number of video spatial layers to use.");
+DEFINE_int(vnum_spatial_layers, 1, "Number of video spatial layers to use.");
 int VideoNumSpatialLayers() {
   return static_cast<int>(FLAG_vnum_spatial_layers);
 }
 
-WEBRTC_DEFINE_int(
-    vinter_layer_pred,
-    2,
-    "Video inter-layer prediction mode. "
-    "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
+DEFINE_int(vinter_layer_pred,
+           2,
+           "Video inter-layer prediction mode. "
+           "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
 InterLayerPredMode VideoInterLayerPred() {
   return IntToInterLayerPredMode(FLAG_vinter_layer_pred);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     vstream0,
     "",
     "Comma separated values describing VideoStream for video stream #0.");
@@ -113,7 +107,7 @@
   return static_cast<std::string>(FLAG_vstream0);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     vstream1,
     "",
     "Comma separated values describing VideoStream for video stream #1.");
@@ -121,7 +115,7 @@
   return static_cast<std::string>(FLAG_vstream1);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     vsl0,
     "",
     "Comma separated values describing SpatialLayer for video layer #0.");
@@ -129,7 +123,7 @@
   return static_cast<std::string>(FLAG_vsl0);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     vsl1,
     "",
     "Comma separated values describing SpatialLayer for video layer #1.");
@@ -137,105 +131,98 @@
   return static_cast<std::string>(FLAG_vsl1);
 }
 
-WEBRTC_DEFINE_int(
-    vselected_tl,
-    -1,
-    "Temporal layer to show or analyze for screenshare. -1 to disable "
-    "filtering.");
+DEFINE_int(vselected_tl,
+           -1,
+           "Temporal layer to show or analyze for screenshare. -1 to disable "
+           "filtering.");
 int VideoSelectedTL() {
   return static_cast<int>(FLAG_vselected_tl);
 }
 
-WEBRTC_DEFINE_int(vselected_stream,
-                  0,
-                  "ID of the stream to show or analyze for screenshare."
-                  "Set to the number of streams to show them all.");
+DEFINE_int(vselected_stream,
+           0,
+           "ID of the stream to show or analyze for screenshare."
+           "Set to the number of streams to show them all.");
 int VideoSelectedStream() {
   return static_cast<int>(FLAG_vselected_stream);
 }
 
-WEBRTC_DEFINE_int(
-    vselected_sl,
-    -1,
-    "Spatial layer to show or analyze for screenshare. -1 to disable "
-    "filtering.");
+DEFINE_int(vselected_sl,
+           -1,
+           "Spatial layer to show or analyze for screenshare. -1 to disable "
+           "filtering.");
 int VideoSelectedSL() {
   return static_cast<int>(FLAG_vselected_sl);
 }
 
 // Flags for screenshare.
-WEBRTC_DEFINE_int(min_transmit_bitrate,
-                  400,
-                  "Min transmit bitrate incl. padding for screenshare.");
+DEFINE_int(min_transmit_bitrate,
+           400,
+           "Min transmit bitrate incl. padding for screenshare.");
 int ScreenshareMinTransmitBitrateKbps() {
   return FLAG_min_transmit_bitrate;
 }
 
-WEBRTC_DEFINE_int(swidth, 1850, "Screenshare width (crops source).");
+DEFINE_int(swidth, 1850, "Screenshare width (crops source).");
 size_t ScreenshareWidth() {
   return static_cast<size_t>(FLAG_swidth);
 }
 
-WEBRTC_DEFINE_int(sheight, 1110, "Screenshare height (crops source).");
+DEFINE_int(sheight, 1110, "Screenshare height (crops source).");
 size_t ScreenshareHeight() {
   return static_cast<size_t>(FLAG_sheight);
 }
 
-WEBRTC_DEFINE_int(sfps, 5, "Frames per second for screenshare.");
+DEFINE_int(sfps, 5, "Frames per second for screenshare.");
 int ScreenshareFps() {
   return static_cast<int>(FLAG_sfps);
 }
 
-WEBRTC_DEFINE_int(starget_bitrate,
-                  100,
-                  "Screenshare stream target bitrate in kbps.");
+DEFINE_int(starget_bitrate, 100, "Screenshare stream target bitrate in kbps.");
 int ScreenshareTargetBitrateKbps() {
   return static_cast<int>(FLAG_starget_bitrate);
 }
 
-WEBRTC_DEFINE_int(smin_bitrate, 100, "Screenshare stream min bitrate in kbps.");
+DEFINE_int(smin_bitrate, 100, "Screenshare stream min bitrate in kbps.");
 int ScreenshareMinBitrateKbps() {
   return static_cast<int>(FLAG_smin_bitrate);
 }
 
-WEBRTC_DEFINE_int(smax_bitrate,
-                  2000,
-                  "Screenshare stream max bitrate in kbps.");
+DEFINE_int(smax_bitrate, 2000, "Screenshare stream max bitrate in kbps.");
 int ScreenshareMaxBitrateKbps() {
   return static_cast<int>(FLAG_smax_bitrate);
 }
 
-WEBRTC_DEFINE_int(snum_temporal_layers,
-                  2,
-                  "Number of temporal layers to use in screenshare.");
+DEFINE_int(snum_temporal_layers,
+           2,
+           "Number of temporal layers to use in screenshare.");
 int ScreenshareNumTemporalLayers() {
   return static_cast<int>(FLAG_snum_temporal_layers);
 }
 
-WEBRTC_DEFINE_int(snum_streams,
-                  0,
-                  "Number of screenshare streams to show or analyze.");
+DEFINE_int(snum_streams,
+           0,
+           "Number of screenshare streams to show or analyze.");
 int ScreenshareNumStreams() {
   return static_cast<int>(FLAG_snum_streams);
 }
 
-WEBRTC_DEFINE_int(snum_spatial_layers,
-                  1,
-                  "Number of screenshare spatial layers to use.");
+DEFINE_int(snum_spatial_layers,
+           1,
+           "Number of screenshare spatial layers to use.");
 int ScreenshareNumSpatialLayers() {
   return static_cast<int>(FLAG_snum_spatial_layers);
 }
 
-WEBRTC_DEFINE_int(
-    sinter_layer_pred,
-    0,
-    "Screenshare inter-layer prediction mode. "
-    "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
+DEFINE_int(sinter_layer_pred,
+           0,
+           "Screenshare inter-layer prediction mode. "
+           "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
 InterLayerPredMode ScreenshareInterLayerPred() {
   return IntToInterLayerPredMode(FLAG_sinter_layer_pred);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     sstream0,
     "",
     "Comma separated values describing VideoStream for screenshare stream #0.");
@@ -243,7 +230,7 @@
   return static_cast<std::string>(FLAG_sstream0);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     sstream1,
     "",
     "Comma separated values describing VideoStream for screenshare stream #1.");
@@ -251,7 +238,7 @@
   return static_cast<std::string>(FLAG_sstream1);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     ssl0,
     "",
     "Comma separated values describing SpatialLayer for screenshare layer #0.");
@@ -259,7 +246,7 @@
   return static_cast<std::string>(FLAG_ssl0);
 }
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     ssl1,
     "",
     "Comma separated values describing SpatialLayer for screenshare layer #1.");
@@ -267,33 +254,31 @@
   return static_cast<std::string>(FLAG_ssl1);
 }
 
-WEBRTC_DEFINE_int(
-    sselected_tl,
-    -1,
-    "Temporal layer to show or analyze for screenshare. -1 to disable "
-    "filtering.");
+DEFINE_int(sselected_tl,
+           -1,
+           "Temporal layer to show or analyze for screenshare. -1 to disable "
+           "filtering.");
 int ScreenshareSelectedTL() {
   return static_cast<int>(FLAG_sselected_tl);
 }
 
-WEBRTC_DEFINE_int(sselected_stream,
-                  0,
-                  "ID of the stream to show or analyze for screenshare."
-                  "Set to the number of streams to show them all.");
+DEFINE_int(sselected_stream,
+           0,
+           "ID of the stream to show or analyze for screenshare."
+           "Set to the number of streams to show them all.");
 int ScreenshareSelectedStream() {
   return static_cast<int>(FLAG_sselected_stream);
 }
 
-WEBRTC_DEFINE_int(
-    sselected_sl,
-    -1,
-    "Spatial layer to show or analyze for screenshare. -1 to disable "
-    "filtering.");
+DEFINE_int(sselected_sl,
+           -1,
+           "Spatial layer to show or analyze for screenshare. -1 to disable "
+           "filtering.");
 int ScreenshareSelectedSL() {
   return static_cast<int>(FLAG_sselected_sl);
 }
 
-WEBRTC_DEFINE_bool(
+DEFINE_bool(
     generate_slides,
     false,
     "Whether to use randomly generated slides or read them from files.");
@@ -301,14 +286,14 @@
   return static_cast<int>(FLAG_generate_slides);
 }
 
-WEBRTC_DEFINE_int(slide_change_interval,
-                  10,
-                  "Interval (in seconds) between simulated slide changes.");
+DEFINE_int(slide_change_interval,
+           10,
+           "Interval (in seconds) between simulated slide changes.");
 int SlideChangeInterval() {
   return static_cast<int>(FLAG_slide_change_interval);
 }
 
-WEBRTC_DEFINE_int(
+DEFINE_int(
     scroll_duration,
     0,
     "Duration (in seconds) during which a slide will be scrolled into place.");
@@ -316,10 +301,9 @@
   return static_cast<int>(FLAG_scroll_duration);
 }
 
-WEBRTC_DEFINE_string(
-    slides,
-    "",
-    "Comma-separated list of *.yuv files to display as slides.");
+DEFINE_string(slides,
+              "",
+              "Comma-separated list of *.yuv files to display as slides.");
 std::vector<std::string> Slides() {
   std::vector<std::string> slides;
   std::string slides_list = FLAG_slides;
@@ -328,31 +312,31 @@
 }
 
 // Flags common with screenshare and video loopback, with equal default values.
-WEBRTC_DEFINE_int(start_bitrate, 600, "Call start bitrate in kbps.");
+DEFINE_int(start_bitrate, 600, "Call start bitrate in kbps.");
 int StartBitrateKbps() {
   return static_cast<int>(FLAG_start_bitrate);
 }
 
-WEBRTC_DEFINE_string(codec, "VP8", "Video codec to use.");
+DEFINE_string(codec, "VP8", "Video codec to use.");
 std::string Codec() {
   return static_cast<std::string>(FLAG_codec);
 }
 
-WEBRTC_DEFINE_bool(analyze_video,
-                   false,
-                   "Analyze video stream (if --duration is present)");
+DEFINE_bool(analyze_video,
+            false,
+            "Analyze video stream (if --duration is present)");
 bool AnalyzeVideo() {
   return static_cast<bool>(FLAG_analyze_video);
 }
 
-WEBRTC_DEFINE_bool(analyze_screenshare,
-                   false,
-                   "Analyze screenshare stream (if --duration is present)");
+DEFINE_bool(analyze_screenshare,
+            false,
+            "Analyze screenshare stream (if --duration is present)");
 bool AnalyzeScreenshare() {
   return static_cast<bool>(FLAG_analyze_screenshare);
 }
 
-WEBRTC_DEFINE_int(
+DEFINE_int(
     duration,
     0,
     "Duration of the test in seconds. If 0, rendered will be shown instead.");
@@ -360,113 +344,100 @@
   return static_cast<int>(FLAG_duration);
 }
 
-WEBRTC_DEFINE_string(output_filename, "", "Target graph data filename.");
+DEFINE_string(output_filename, "", "Target graph data filename.");
 std::string OutputFilename() {
   return static_cast<std::string>(FLAG_output_filename);
 }
 
-WEBRTC_DEFINE_string(graph_title,
-                     "",
-                     "If empty, title will be generated automatically.");
+DEFINE_string(graph_title,
+              "",
+              "If empty, title will be generated automatically.");
 std::string GraphTitle() {
   return static_cast<std::string>(FLAG_graph_title);
 }
 
-WEBRTC_DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
+DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
 int LossPercent() {
   return static_cast<int>(FLAG_loss_percent);
 }
 
-WEBRTC_DEFINE_int(avg_burst_loss_length,
-                  -1,
-                  "Average burst length of lost packets.");
+DEFINE_int(avg_burst_loss_length, -1, "Average burst length of lost packets.");
 int AvgBurstLossLength() {
   return static_cast<int>(FLAG_avg_burst_loss_length);
 }
 
-WEBRTC_DEFINE_int(link_capacity,
-                  0,
-                  "Capacity (kbps) of the fake link. 0 means infinite.");
+DEFINE_int(link_capacity,
+           0,
+           "Capacity (kbps) of the fake link. 0 means infinite.");
 int LinkCapacityKbps() {
   return static_cast<int>(FLAG_link_capacity);
 }
 
-WEBRTC_DEFINE_int(queue_size,
-                  0,
-                  "Size of the bottleneck link queue in packets.");
+DEFINE_int(queue_size, 0, "Size of the bottleneck link queue in packets.");
 int QueueSize() {
   return static_cast<int>(FLAG_queue_size);
 }
 
-WEBRTC_DEFINE_int(avg_propagation_delay_ms,
-                  0,
-                  "Average link propagation delay in ms.");
+DEFINE_int(avg_propagation_delay_ms,
+           0,
+           "Average link propagation delay in ms.");
 int AvgPropagationDelayMs() {
   return static_cast<int>(FLAG_avg_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_string(rtc_event_log_name,
-                     "",
-                     "Filename for rtc event log. Two files "
-                     "with \"_send\" and \"_recv\" suffixes will be created. "
-                     "Works only when --duration is set.");
+DEFINE_string(rtc_event_log_name,
+              "",
+              "Filename for rtc event log. Two files "
+              "with \"_send\" and \"_recv\" suffixes will be created. "
+              "Works only when --duration is set.");
 std::string RtcEventLogName() {
   return static_cast<std::string>(FLAG_rtc_event_log_name);
 }
 
-WEBRTC_DEFINE_string(rtp_dump_name,
-                     "",
-                     "Filename for dumped received RTP stream.");
+DEFINE_string(rtp_dump_name, "", "Filename for dumped received RTP stream.");
 std::string RtpDumpName() {
   return static_cast<std::string>(FLAG_rtp_dump_name);
 }
 
-WEBRTC_DEFINE_int(std_propagation_delay_ms,
-                  0,
-                  "Link propagation delay standard deviation in ms.");
+DEFINE_int(std_propagation_delay_ms,
+           0,
+           "Link propagation delay standard deviation in ms.");
 int StdPropagationDelayMs() {
   return static_cast<int>(FLAG_std_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_string(
-    encoded_frame_path,
-    "",
-    "The base path for encoded frame logs. Created files will have "
-    "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
+DEFINE_string(encoded_frame_path,
+              "",
+              "The base path for encoded frame logs. Created files will have "
+              "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
 std::string EncodedFramePath() {
   return static_cast<std::string>(FLAG_encoded_frame_path);
 }
 
-WEBRTC_DEFINE_bool(logs, false, "print logs to stderr");
+DEFINE_bool(logs, false, "print logs to stderr");
 
-WEBRTC_DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
+DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
 
-WEBRTC_DEFINE_bool(generic_descriptor,
-                   false,
-                   "Use the generic frame descriptor.");
+DEFINE_bool(generic_descriptor, false, "Use the generic frame descriptor.");
 
-WEBRTC_DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
+DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
 
-WEBRTC_DEFINE_bool(use_ulpfec,
-                   false,
-                   "Use RED+ULPFEC forward error correction.");
+DEFINE_bool(use_ulpfec, false, "Use RED+ULPFEC forward error correction.");
 
-WEBRTC_DEFINE_bool(use_flexfec, false, "Use FlexFEC forward error correction.");
+DEFINE_bool(use_flexfec, false, "Use FlexFEC forward error correction.");
 
-WEBRTC_DEFINE_bool(audio, false, "Add audio stream");
+DEFINE_bool(audio, false, "Add audio stream");
 
-WEBRTC_DEFINE_bool(audio_video_sync,
-                   false,
-                   "Sync audio and video stream (no effect if"
-                   " audio is false)");
+DEFINE_bool(audio_video_sync,
+            false,
+            "Sync audio and video stream (no effect if"
+            " audio is false)");
 
-WEBRTC_DEFINE_bool(audio_dtx,
-                   false,
-                   "Enable audio DTX (no effect if audio is false)");
+DEFINE_bool(audio_dtx, false, "Enable audio DTX (no effect if audio is false)");
 
-WEBRTC_DEFINE_bool(video, true, "Add video stream");
+DEFINE_bool(video, true, "Add video stream");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
@@ -475,16 +446,15 @@
     "trials are separated by \"/\"");
 
 // Video-specific flags.
-WEBRTC_DEFINE_string(
-    vclip,
-    "",
-    "Name of the clip to show. If empty, the camera is used. Use "
-    "\"Generator\" for chroma generator.");
+DEFINE_string(vclip,
+              "",
+              "Name of the clip to show. If empty, the camera is used. Use "
+              "\"Generator\" for chroma generator.");
 std::string VideoClip() {
   return static_cast<std::string>(FLAG_vclip);
 }
 
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_bool(help, false, "prints this message");
 
 }  // namespace flags
 
diff --git a/video/video_analyzer.cc b/video/video_analyzer.cc
index 6d16b1a..1b1c557 100644
--- a/video/video_analyzer.cc
+++ b/video/video_analyzer.cc
@@ -25,11 +25,10 @@
 #include "test/testsupport/perf_test.h"
 #include "test/testsupport/test_artifacts.h"
 
-WEBRTC_DEFINE_bool(
-    save_worst_frame,
-    false,
-    "Enable saving a frame with the lowest PSNR to a jpeg file in the "
-    "test_artifacts_dir");
+DEFINE_bool(save_worst_frame,
+            false,
+            "Enable saving a frame with the lowest PSNR to a jpeg file in the "
+            "test_artifacts_dir");
 
 namespace webrtc {
 namespace {
diff --git a/video/video_loopback.cc b/video/video_loopback.cc
index 4cdddb9..b6715ac 100644
--- a/video/video_loopback.cc
+++ b/video/video_loopback.cc
@@ -22,62 +22,61 @@
 namespace flags {
 
 // Flags common with screenshare loopback, with different default values.
-WEBRTC_DEFINE_int(width, 640, "Video width.");
+DEFINE_int(width, 640, "Video width.");
 size_t Width() {
   return static_cast<size_t>(FLAG_width);
 }
 
-WEBRTC_DEFINE_int(height, 480, "Video height.");
+DEFINE_int(height, 480, "Video height.");
 size_t Height() {
   return static_cast<size_t>(FLAG_height);
 }
 
-WEBRTC_DEFINE_int(fps, 30, "Frames per second.");
+DEFINE_int(fps, 30, "Frames per second.");
 int Fps() {
   return static_cast<int>(FLAG_fps);
 }
 
-WEBRTC_DEFINE_int(capture_device_index, 0, "Capture device to select");
+DEFINE_int(capture_device_index, 0, "Capture device to select");
 size_t GetCaptureDevice() {
   return static_cast<size_t>(FLAG_capture_device_index);
 }
 
-WEBRTC_DEFINE_int(min_bitrate, 50, "Call and stream min bitrate in kbps.");
+DEFINE_int(min_bitrate, 50, "Call and stream min bitrate in kbps.");
 int MinBitrateKbps() {
   return static_cast<int>(FLAG_min_bitrate);
 }
 
-WEBRTC_DEFINE_int(start_bitrate, 300, "Call start bitrate in kbps.");
+DEFINE_int(start_bitrate, 300, "Call start bitrate in kbps.");
 int StartBitrateKbps() {
   return static_cast<int>(FLAG_start_bitrate);
 }
 
-WEBRTC_DEFINE_int(target_bitrate, 800, "Stream target bitrate in kbps.");
+DEFINE_int(target_bitrate, 800, "Stream target bitrate in kbps.");
 int TargetBitrateKbps() {
   return static_cast<int>(FLAG_target_bitrate);
 }
 
-WEBRTC_DEFINE_int(max_bitrate, 800, "Call and stream max bitrate in kbps.");
+DEFINE_int(max_bitrate, 800, "Call and stream max bitrate in kbps.");
 int MaxBitrateKbps() {
   return static_cast<int>(FLAG_max_bitrate);
 }
 
-WEBRTC_DEFINE_bool(suspend_below_min_bitrate,
-                   false,
-                   "Suspends video below the configured min bitrate.");
+DEFINE_bool(suspend_below_min_bitrate,
+            false,
+            "Suspends video below the configured min bitrate.");
 
-WEBRTC_DEFINE_int(num_temporal_layers,
-                  1,
-                  "Number of temporal layers. Set to 1-4 to override.");
+DEFINE_int(num_temporal_layers,
+           1,
+           "Number of temporal layers. Set to 1-4 to override.");
 int NumTemporalLayers() {
   return static_cast<int>(FLAG_num_temporal_layers);
 }
 
-WEBRTC_DEFINE_int(
-    inter_layer_pred,
-    2,
-    "Inter-layer prediction mode. "
-    "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
+DEFINE_int(inter_layer_pred,
+           2,
+           "Inter-layer prediction mode. "
+           "0 - enabled, 1 - disabled, 2 - enabled only for key pictures.");
 InterLayerPredMode InterLayerPred() {
   if (FLAG_inter_layer_pred == 0) {
     return InterLayerPredMode::kOn;
@@ -90,20 +89,19 @@
 }
 
 // Flags common with screenshare loopback, with equal default values.
-WEBRTC_DEFINE_string(codec, "VP8", "Video codec to use.");
+DEFINE_string(codec, "VP8", "Video codec to use.");
 std::string Codec() {
   return static_cast<std::string>(FLAG_codec);
 }
 
-WEBRTC_DEFINE_int(
-    selected_tl,
-    -1,
-    "Temporal layer to show or analyze. -1 to disable filtering.");
+DEFINE_int(selected_tl,
+           -1,
+           "Temporal layer to show or analyze. -1 to disable filtering.");
 int SelectedTL() {
   return static_cast<int>(FLAG_selected_tl);
 }
 
-WEBRTC_DEFINE_int(
+DEFINE_int(
     duration,
     0,
     "Duration of the test in seconds. If 0, rendered will be shown instead.");
@@ -111,174 +109,156 @@
   return static_cast<int>(FLAG_duration);
 }
 
-WEBRTC_DEFINE_string(output_filename, "", "Target graph data filename.");
+DEFINE_string(output_filename, "", "Target graph data filename.");
 std::string OutputFilename() {
   return static_cast<std::string>(FLAG_output_filename);
 }
 
-WEBRTC_DEFINE_string(graph_title,
-                     "",
-                     "If empty, title will be generated automatically.");
+DEFINE_string(graph_title,
+              "",
+              "If empty, title will be generated automatically.");
 std::string GraphTitle() {
   return static_cast<std::string>(FLAG_graph_title);
 }
 
-WEBRTC_DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
+DEFINE_int(loss_percent, 0, "Percentage of packets randomly lost.");
 int LossPercent() {
   return static_cast<int>(FLAG_loss_percent);
 }
 
-WEBRTC_DEFINE_int(avg_burst_loss_length,
-                  -1,
-                  "Average burst length of lost packets.");
+DEFINE_int(avg_burst_loss_length, -1, "Average burst length of lost packets.");
 int AvgBurstLossLength() {
   return static_cast<int>(FLAG_avg_burst_loss_length);
 }
 
-WEBRTC_DEFINE_int(link_capacity,
-                  0,
-                  "Capacity (kbps) of the fake link. 0 means infinite.");
+DEFINE_int(link_capacity,
+           0,
+           "Capacity (kbps) of the fake link. 0 means infinite.");
 int LinkCapacityKbps() {
   return static_cast<int>(FLAG_link_capacity);
 }
 
-WEBRTC_DEFINE_int(queue_size,
-                  0,
-                  "Size of the bottleneck link queue in packets.");
+DEFINE_int(queue_size, 0, "Size of the bottleneck link queue in packets.");
 int QueueSize() {
   return static_cast<int>(FLAG_queue_size);
 }
 
-WEBRTC_DEFINE_int(avg_propagation_delay_ms,
-                  0,
-                  "Average link propagation delay in ms.");
+DEFINE_int(avg_propagation_delay_ms,
+           0,
+           "Average link propagation delay in ms.");
 int AvgPropagationDelayMs() {
   return static_cast<int>(FLAG_avg_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_string(rtc_event_log_name,
-                     "",
-                     "Filename for rtc event log. Two files "
-                     "with \"_send\" and \"_recv\" suffixes will be created.");
+DEFINE_string(rtc_event_log_name,
+              "",
+              "Filename for rtc event log. Two files "
+              "with \"_send\" and \"_recv\" suffixes will be created.");
 std::string RtcEventLogName() {
   return static_cast<std::string>(FLAG_rtc_event_log_name);
 }
 
-WEBRTC_DEFINE_string(rtp_dump_name,
-                     "",
-                     "Filename for dumped received RTP stream.");
+DEFINE_string(rtp_dump_name, "", "Filename for dumped received RTP stream.");
 std::string RtpDumpName() {
   return static_cast<std::string>(FLAG_rtp_dump_name);
 }
 
-WEBRTC_DEFINE_int(std_propagation_delay_ms,
-                  0,
-                  "Link propagation delay standard deviation in ms.");
+DEFINE_int(std_propagation_delay_ms,
+           0,
+           "Link propagation delay standard deviation in ms.");
 int StdPropagationDelayMs() {
   return static_cast<int>(FLAG_std_propagation_delay_ms);
 }
 
-WEBRTC_DEFINE_int(num_streams, 0, "Number of streams to show or analyze.");
+DEFINE_int(num_streams, 0, "Number of streams to show or analyze.");
 int NumStreams() {
   return static_cast<int>(FLAG_num_streams);
 }
 
-WEBRTC_DEFINE_int(selected_stream,
-                  0,
-                  "ID of the stream to show or analyze. "
-                  "Set to the number of streams to show them all.");
+DEFINE_int(selected_stream,
+           0,
+           "ID of the stream to show or analyze. "
+           "Set to the number of streams to show them all.");
 int SelectedStream() {
   return static_cast<int>(FLAG_selected_stream);
 }
 
-WEBRTC_DEFINE_int(num_spatial_layers, 1, "Number of spatial layers to use.");
+DEFINE_int(num_spatial_layers, 1, "Number of spatial layers to use.");
 int NumSpatialLayers() {
   return static_cast<int>(FLAG_num_spatial_layers);
 }
 
-WEBRTC_DEFINE_int(selected_sl,
-                  -1,
-                  "Spatial layer to show or analyze. -1 to disable filtering.");
+DEFINE_int(selected_sl,
+           -1,
+           "Spatial layer to show or analyze. -1 to disable filtering.");
 int SelectedSL() {
   return static_cast<int>(FLAG_selected_sl);
 }
 
-WEBRTC_DEFINE_string(
-    stream0,
-    "",
-    "Comma separated values describing VideoStream for stream #0.");
+DEFINE_string(stream0,
+              "",
+              "Comma separated values describing VideoStream for stream #0.");
 std::string Stream0() {
   return static_cast<std::string>(FLAG_stream0);
 }
 
-WEBRTC_DEFINE_string(
-    stream1,
-    "",
-    "Comma separated values describing VideoStream for stream #1.");
+DEFINE_string(stream1,
+              "",
+              "Comma separated values describing VideoStream for stream #1.");
 std::string Stream1() {
   return static_cast<std::string>(FLAG_stream1);
 }
 
-WEBRTC_DEFINE_string(
-    sl0,
-    "",
-    "Comma separated values describing SpatialLayer for layer #0.");
+DEFINE_string(sl0,
+              "",
+              "Comma separated values describing SpatialLayer for layer #0.");
 std::string SL0() {
   return static_cast<std::string>(FLAG_sl0);
 }
 
-WEBRTC_DEFINE_string(
-    sl1,
-    "",
-    "Comma separated values describing SpatialLayer for layer #1.");
+DEFINE_string(sl1,
+              "",
+              "Comma separated values describing SpatialLayer for layer #1.");
 std::string SL1() {
   return static_cast<std::string>(FLAG_sl1);
 }
 
-WEBRTC_DEFINE_string(
-    encoded_frame_path,
-    "",
-    "The base path for encoded frame logs. Created files will have "
-    "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
+DEFINE_string(encoded_frame_path,
+              "",
+              "The base path for encoded frame logs. Created files will have "
+              "the form <encoded_frame_path>.<n>.(recv|send.<m>).ivf");
 std::string EncodedFramePath() {
   return static_cast<std::string>(FLAG_encoded_frame_path);
 }
 
-WEBRTC_DEFINE_bool(logs, false, "print logs to stderr");
+DEFINE_bool(logs, false, "print logs to stderr");
 
-WEBRTC_DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
+DEFINE_bool(send_side_bwe, true, "Use send-side bandwidth estimation");
 
-WEBRTC_DEFINE_bool(generic_descriptor,
-                   false,
-                   "Use the generic frame descriptor.");
+DEFINE_bool(generic_descriptor, false, "Use the generic frame descriptor.");
 
-WEBRTC_DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
+DEFINE_bool(allow_reordering, false, "Allow packet reordering to occur");
 
-WEBRTC_DEFINE_bool(use_ulpfec,
-                   false,
-                   "Use RED+ULPFEC forward error correction.");
+DEFINE_bool(use_ulpfec, false, "Use RED+ULPFEC forward error correction.");
 
-WEBRTC_DEFINE_bool(use_flexfec, false, "Use FlexFEC forward error correction.");
+DEFINE_bool(use_flexfec, false, "Use FlexFEC forward error correction.");
 
-WEBRTC_DEFINE_bool(audio, false, "Add audio stream");
+DEFINE_bool(audio, false, "Add audio stream");
 
-WEBRTC_DEFINE_bool(
-    use_real_adm,
-    false,
-    "Use real ADM instead of fake (no effect if audio is false)");
+DEFINE_bool(use_real_adm,
+            false,
+            "Use real ADM instead of fake (no effect if audio is false)");
 
-WEBRTC_DEFINE_bool(audio_video_sync,
-                   false,
-                   "Sync audio and video stream (no effect if"
-                   " audio is false)");
+DEFINE_bool(audio_video_sync,
+            false,
+            "Sync audio and video stream (no effect if"
+            " audio is false)");
 
-WEBRTC_DEFINE_bool(audio_dtx,
-                   false,
-                   "Enable audio DTX (no effect if audio is false)");
+DEFINE_bool(audio_dtx, false, "Enable audio DTX (no effect if audio is false)");
 
-WEBRTC_DEFINE_bool(video, true, "Add video stream");
+DEFINE_bool(video, true, "Add video stream");
 
-WEBRTC_DEFINE_string(
+DEFINE_string(
     force_fieldtrials,
     "",
     "Field trials control experimental feature code which can be forced. "
@@ -287,15 +267,14 @@
     "trials are separated by \"/\"");
 
 // Video-specific flags.
-WEBRTC_DEFINE_string(
-    clip,
-    "",
-    "Name of the clip to show. If empty, using chroma generator.");
+DEFINE_string(clip,
+              "",
+              "Name of the clip to show. If empty, using chroma generator.");
 std::string Clip() {
   return static_cast<std::string>(FLAG_clip);
 }
 
-WEBRTC_DEFINE_bool(help, false, "prints this message");
+DEFINE_bool(help, false, "prints this message");
 
 }  // namespace flags