| // Protocol Buffers - Google's data interchange format |
| // Copyright 2008 Google Inc. All rights reserved. |
| // https://developers.google.com/protocol-buffers/ |
| // |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: |
| // |
| // * Redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above |
| // copyright notice, this list of conditions and the following disclaimer |
| // in the documentation and/or other materials provided with the |
| // distribution. |
| // * Neither the name of Google Inc. nor the names of its |
| // contributors may be used to endorse or promote products derived from |
| // this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #include <stdarg.h> |
| #include <string> |
| #include <fstream> |
| |
| #include "conformance.pb.h" |
| #include "conformance_test.h" |
| #include <google/protobuf/test_messages_proto3.pb.h> |
| |
| #include <google/protobuf/stubs/common.h> |
| #include <google/protobuf/stubs/stringprintf.h> |
| #include <google/protobuf/text_format.h> |
| #include <google/protobuf/util/field_comparator.h> |
| #include <google/protobuf/util/json_util.h> |
| #include <google/protobuf/util/message_differencer.h> |
| #include <google/protobuf/util/type_resolver_util.h> |
| #include <google/protobuf/wire_format_lite.h> |
| |
| #include "third_party/jsoncpp/json.h" |
| |
| using conformance::ConformanceRequest; |
| using conformance::ConformanceResponse; |
| using conformance::WireFormat; |
| using google::protobuf::Descriptor; |
| using google::protobuf::FieldDescriptor; |
| using google::protobuf::internal::WireFormatLite; |
| using google::protobuf::TextFormat; |
| using google::protobuf::util::DefaultFieldComparator; |
| using google::protobuf::util::JsonToBinaryString; |
| using google::protobuf::util::MessageDifferencer; |
| using google::protobuf::util::NewTypeResolverForDescriptorPool; |
| using google::protobuf::util::Status; |
| using protobuf_test_messages::proto3::TestAllTypes; |
| using std::string; |
| |
| namespace { |
| |
| static const char kTypeUrlPrefix[] = "type.googleapis.com"; |
| |
| static string GetTypeUrl(const Descriptor* message) { |
| return string(kTypeUrlPrefix) + "/" + message->full_name(); |
| } |
| |
| /* Routines for building arbitrary protos *************************************/ |
| |
| // We would use CodedOutputStream except that we want more freedom to build |
| // arbitrary protos (even invalid ones). |
| |
| const string empty; |
| |
| string cat(const string& a, const string& b, |
| const string& c = empty, |
| const string& d = empty, |
| const string& e = empty, |
| const string& f = empty, |
| const string& g = empty, |
| const string& h = empty, |
| const string& i = empty, |
| const string& j = empty, |
| const string& k = empty, |
| const string& l = empty) { |
| string ret; |
| ret.reserve(a.size() + b.size() + c.size() + d.size() + e.size() + f.size() + |
| g.size() + h.size() + i.size() + j.size() + k.size() + l.size()); |
| ret.append(a); |
| ret.append(b); |
| ret.append(c); |
| ret.append(d); |
| ret.append(e); |
| ret.append(f); |
| ret.append(g); |
| ret.append(h); |
| ret.append(i); |
| ret.append(j); |
| ret.append(k); |
| ret.append(l); |
| return ret; |
| } |
| |
| // The maximum number of bytes that it takes to encode a 64-bit varint. |
| #define VARINT_MAX_LEN 10 |
| |
| size_t vencode64(uint64_t val, int over_encoded_bytes, char *buf) { |
| if (val == 0) { buf[0] = 0; return 1; } |
| size_t i = 0; |
| while (val) { |
| uint8_t byte = val & 0x7fU; |
| val >>= 7; |
| if (val || over_encoded_bytes) byte |= 0x80U; |
| buf[i++] = byte; |
| } |
| while (over_encoded_bytes--) { |
| assert(i < 10); |
| uint8_t byte = over_encoded_bytes ? 0x80 : 0; |
| buf[i++] = byte; |
| } |
| return i; |
| } |
| |
| string varint(uint64_t x) { |
| char buf[VARINT_MAX_LEN]; |
| size_t len = vencode64(x, 0, buf); |
| return string(buf, len); |
| } |
| |
| // Encodes a varint that is |extra| bytes longer than it needs to be, but still |
| // valid. |
| string longvarint(uint64_t x, int extra) { |
| char buf[VARINT_MAX_LEN]; |
| size_t len = vencode64(x, extra, buf); |
| return string(buf, len); |
| } |
| |
| // TODO: proper byte-swapping for big-endian machines. |
| string fixed32(void *data) { return string(static_cast<char*>(data), 4); } |
| string fixed64(void *data) { return string(static_cast<char*>(data), 8); } |
| |
| string delim(const string& buf) { return cat(varint(buf.size()), buf); } |
| string u32(uint32_t u32) { return fixed32(&u32); } |
| string u64(uint64_t u64) { return fixed64(&u64); } |
| string flt(float f) { return fixed32(&f); } |
| string dbl(double d) { return fixed64(&d); } |
| string zz32(int32_t x) { return varint(WireFormatLite::ZigZagEncode32(x)); } |
| string zz64(int64_t x) { return varint(WireFormatLite::ZigZagEncode64(x)); } |
| |
| string tag(uint32_t fieldnum, char wire_type) { |
| return varint((fieldnum << 3) | wire_type); |
| } |
| |
| string submsg(uint32_t fn, const string& buf) { |
| return cat( tag(fn, WireFormatLite::WIRETYPE_LENGTH_DELIMITED), delim(buf) ); |
| } |
| |
| #define UNKNOWN_FIELD 666 |
| |
| const FieldDescriptor* GetFieldForType(FieldDescriptor::Type type, |
| bool repeated) { |
| const Descriptor* d = TestAllTypes().GetDescriptor(); |
| for (int i = 0; i < d->field_count(); i++) { |
| const FieldDescriptor* f = d->field(i); |
| if (f->type() == type && f->is_repeated() == repeated) { |
| return f; |
| } |
| } |
| GOOGLE_LOG(FATAL) << "Couldn't find field with type " << (int)type; |
| return nullptr; |
| } |
| |
| string UpperCase(string str) { |
| for (int i = 0; i < str.size(); i++) { |
| str[i] = toupper(str[i]); |
| } |
| return str; |
| } |
| |
| } // anonymous namespace |
| |
| namespace google { |
| namespace protobuf { |
| |
| void ConformanceTestSuite::ReportSuccess(const string& test_name) { |
| if (expected_to_fail_.erase(test_name) != 0) { |
| StringAppendF(&output_, |
| "ERROR: test %s is in the failure list, but test succeeded. " |
| "Remove it from the failure list.\n", |
| test_name.c_str()); |
| unexpected_succeeding_tests_.insert(test_name); |
| } |
| successes_++; |
| } |
| |
| void ConformanceTestSuite::ReportFailure(const string& test_name, |
| ConformanceLevel level, |
| const ConformanceRequest& request, |
| const ConformanceResponse& response, |
| const char* fmt, ...) { |
| if (expected_to_fail_.erase(test_name) == 1) { |
| expected_failures_++; |
| if (!verbose_) |
| return; |
| } else if (level == RECOMMENDED && !enforce_recommended_) { |
| StringAppendF(&output_, "WARNING, test=%s: ", test_name.c_str()); |
| } else { |
| StringAppendF(&output_, "ERROR, test=%s: ", test_name.c_str()); |
| unexpected_failing_tests_.insert(test_name); |
| } |
| va_list args; |
| va_start(args, fmt); |
| StringAppendV(&output_, fmt, args); |
| va_end(args); |
| StringAppendF(&output_, " request=%s, response=%s\n", |
| request.ShortDebugString().c_str(), |
| response.ShortDebugString().c_str()); |
| } |
| |
| void ConformanceTestSuite::ReportSkip(const string& test_name, |
| const ConformanceRequest& request, |
| const ConformanceResponse& response) { |
| if (verbose_) { |
| StringAppendF(&output_, "SKIPPED, test=%s request=%s, response=%s\n", |
| test_name.c_str(), request.ShortDebugString().c_str(), |
| response.ShortDebugString().c_str()); |
| } |
| skipped_.insert(test_name); |
| } |
| |
| string ConformanceTestSuite::ConformanceLevelToString(ConformanceLevel level) { |
| switch (level) { |
| case REQUIRED: return "Required"; |
| case RECOMMENDED: return "Recommended"; |
| } |
| GOOGLE_LOG(FATAL) << "Unknown value: " << level; |
| return ""; |
| } |
| |
| void ConformanceTestSuite::RunTest(const string& test_name, |
| const ConformanceRequest& request, |
| ConformanceResponse* response) { |
| if (test_names_.insert(test_name).second == false) { |
| GOOGLE_LOG(FATAL) << "Duplicated test name: " << test_name; |
| } |
| |
| string serialized_request; |
| string serialized_response; |
| request.SerializeToString(&serialized_request); |
| |
| runner_->RunTest(test_name, serialized_request, &serialized_response); |
| |
| if (!response->ParseFromString(serialized_response)) { |
| response->Clear(); |
| response->set_runtime_error("response proto could not be parsed."); |
| } |
| |
| if (verbose_) { |
| StringAppendF(&output_, "conformance test: name=%s, request=%s, response=%s\n", |
| test_name.c_str(), |
| request.ShortDebugString().c_str(), |
| response->ShortDebugString().c_str()); |
| } |
| } |
| |
| void ConformanceTestSuite::RunValidInputTest( |
| const string& test_name, ConformanceLevel level, const string& input, |
| WireFormat input_format, const string& equivalent_text_format, |
| WireFormat requested_output) { |
| TestAllTypes reference_message; |
| GOOGLE_CHECK( |
| TextFormat::ParseFromString(equivalent_text_format, &reference_message)) |
| << "Failed to parse data for test case: " << test_name |
| << ", data: " << equivalent_text_format; |
| |
| ConformanceRequest request; |
| ConformanceResponse response; |
| |
| switch (input_format) { |
| case conformance::PROTOBUF: |
| request.set_protobuf_payload(input); |
| break; |
| |
| case conformance::JSON: |
| request.set_json_payload(input); |
| break; |
| |
| default: |
| GOOGLE_LOG(FATAL) << "Unspecified input format"; |
| } |
| |
| request.set_requested_output_format(requested_output); |
| |
| RunTest(test_name, request, &response); |
| |
| TestAllTypes test_message; |
| |
| switch (response.result_case()) { |
| case ConformanceResponse::RESULT_NOT_SET: |
| ReportFailure(test_name, level, request, response, |
| "Response didn't have any field in the Response."); |
| return; |
| |
| case ConformanceResponse::kParseError: |
| case ConformanceResponse::kRuntimeError: |
| case ConformanceResponse::kSerializeError: |
| ReportFailure(test_name, level, request, response, |
| "Failed to parse input or produce output."); |
| return; |
| |
| case ConformanceResponse::kSkipped: |
| ReportSkip(test_name, request, response); |
| return; |
| |
| case ConformanceResponse::kJsonPayload: { |
| if (requested_output != conformance::JSON) { |
| ReportFailure( |
| test_name, level, request, response, |
| "Test was asked for protobuf output but provided JSON instead."); |
| return; |
| } |
| string binary_protobuf; |
| Status status = |
| JsonToBinaryString(type_resolver_.get(), type_url_, |
| response.json_payload(), &binary_protobuf); |
| if (!status.ok()) { |
| ReportFailure(test_name, level, request, response, |
| "JSON output we received from test was unparseable."); |
| return; |
| } |
| |
| if (!test_message.ParseFromString(binary_protobuf)) { |
| ReportFailure(test_name, level, request, response, |
| "INTERNAL ERROR: internal JSON->protobuf transcode " |
| "yielded unparseable proto."); |
| return; |
| } |
| |
| break; |
| } |
| |
| case ConformanceResponse::kProtobufPayload: { |
| if (requested_output != conformance::PROTOBUF) { |
| ReportFailure( |
| test_name, level, request, response, |
| "Test was asked for JSON output but provided protobuf instead."); |
| return; |
| } |
| |
| if (!test_message.ParseFromString(response.protobuf_payload())) { |
| ReportFailure(test_name, level, request, response, |
| "Protobuf output we received from test was unparseable."); |
| return; |
| } |
| |
| break; |
| } |
| |
| default: |
| GOOGLE_LOG(FATAL) << test_name << ": unknown payload type: " |
| << response.result_case(); |
| } |
| |
| MessageDifferencer differencer; |
| DefaultFieldComparator field_comparator; |
| field_comparator.set_treat_nan_as_equal(true); |
| differencer.set_field_comparator(&field_comparator); |
| string differences; |
| differencer.ReportDifferencesToString(&differences); |
| |
| if (differencer.Compare(reference_message, test_message)) { |
| ReportSuccess(test_name); |
| } else { |
| ReportFailure(test_name, level, request, response, |
| "Output was not equivalent to reference message: %s.", |
| differences.c_str()); |
| } |
| } |
| |
| // Expect that this precise protobuf will cause a parse error. |
| void ConformanceTestSuite::ExpectParseFailureForProto( |
| const string& proto, const string& test_name, ConformanceLevel level) { |
| ConformanceRequest request; |
| ConformanceResponse response; |
| request.set_protobuf_payload(proto); |
| string effective_test_name = ConformanceLevelToString(level) + |
| ".ProtobufInput." + test_name; |
| |
| // We don't expect output, but if the program erroneously accepts the protobuf |
| // we let it send its response as this. We must not leave it unspecified. |
| request.set_requested_output_format(conformance::PROTOBUF); |
| |
| RunTest(effective_test_name, request, &response); |
| if (response.result_case() == ConformanceResponse::kParseError) { |
| ReportSuccess(effective_test_name); |
| } else if (response.result_case() == ConformanceResponse::kSkipped) { |
| ReportSkip(effective_test_name, request, response); |
| } else { |
| ReportFailure(effective_test_name, level, request, response, |
| "Should have failed to parse, but didn't."); |
| } |
| } |
| |
| // Expect that this protobuf will cause a parse error, even if it is followed |
| // by valid protobuf data. We can try running this twice: once with this |
| // data verbatim and once with this data followed by some valid data. |
| // |
| // TODO(haberman): implement the second of these. |
| void ConformanceTestSuite::ExpectHardParseFailureForProto( |
| const string& proto, const string& test_name, ConformanceLevel level) { |
| return ExpectParseFailureForProto(proto, test_name, level); |
| } |
| |
| void ConformanceTestSuite::RunValidJsonTest( |
| const string& test_name, ConformanceLevel level, const string& input_json, |
| const string& equivalent_text_format) { |
| RunValidInputTest( |
| ConformanceLevelToString(level) + ".JsonInput." + test_name + |
| ".ProtobufOutput", level, input_json, conformance::JSON, |
| equivalent_text_format, conformance::PROTOBUF); |
| RunValidInputTest( |
| ConformanceLevelToString(level) + ".JsonInput." + test_name + |
| ".JsonOutput", level, input_json, conformance::JSON, |
| equivalent_text_format, conformance::JSON); |
| } |
| |
| void ConformanceTestSuite::RunValidJsonTestWithProtobufInput( |
| const string& test_name, ConformanceLevel level, const TestAllTypes& input, |
| const string& equivalent_text_format) { |
| RunValidInputTest( |
| ConformanceLevelToString(level) + ".ProtobufInput." + test_name + |
| ".JsonOutput", level, input.SerializeAsString(), conformance::PROTOBUF, |
| equivalent_text_format, conformance::JSON); |
| } |
| |
| void ConformanceTestSuite::RunValidProtobufTest( |
| const string& test_name, ConformanceLevel level, |
| const string& input_protobuf, const string& equivalent_text_format) { |
| RunValidInputTest( |
| ConformanceLevelToString(level) + ".ProtobufInput." + test_name + |
| ".ProtobufOutput", level, input_protobuf, conformance::PROTOBUF, |
| equivalent_text_format, conformance::PROTOBUF); |
| RunValidInputTest( |
| ConformanceLevelToString(level) + ".ProtobufInput." + test_name + |
| ".JsonOutput", level, input_protobuf, conformance::PROTOBUF, |
| equivalent_text_format, conformance::JSON); |
| } |
| |
| void ConformanceTestSuite::RunValidProtobufTestWithMessage( |
| const string& test_name, ConformanceLevel level, const TestAllTypes& input, |
| const string& equivalent_text_format) { |
| RunValidProtobufTest(test_name, level, input.SerializeAsString(), equivalent_text_format); |
| } |
| |
| // According to proto3 JSON specification, JSON serializers follow more strict |
| // rules than parsers (e.g., a serializer must serialize int32 values as JSON |
| // numbers while the parser is allowed to accept them as JSON strings). This |
| // method allows strict checking on a proto3 JSON serializer by inspecting |
| // the JSON output directly. |
| void ConformanceTestSuite::RunValidJsonTestWithValidator( |
| const string& test_name, ConformanceLevel level, const string& input_json, |
| const Validator& validator) { |
| ConformanceRequest request; |
| ConformanceResponse response; |
| request.set_json_payload(input_json); |
| request.set_requested_output_format(conformance::JSON); |
| |
| string effective_test_name = ConformanceLevelToString(level) + |
| ".JsonInput." + test_name + ".Validator"; |
| |
| RunTest(effective_test_name, request, &response); |
| |
| if (response.result_case() == ConformanceResponse::kSkipped) { |
| ReportSkip(effective_test_name, request, response); |
| return; |
| } |
| |
| if (response.result_case() != ConformanceResponse::kJsonPayload) { |
| ReportFailure(effective_test_name, level, request, response, |
| "Expected JSON payload but got type %d.", |
| response.result_case()); |
| return; |
| } |
| Json::Reader reader; |
| Json::Value value; |
| if (!reader.parse(response.json_payload(), value)) { |
| ReportFailure(effective_test_name, level, request, response, |
| "JSON payload cannot be parsed as valid JSON: %s", |
| reader.getFormattedErrorMessages().c_str()); |
| return; |
| } |
| if (!validator(value)) { |
| ReportFailure(effective_test_name, level, request, response, |
| "JSON payload validation failed."); |
| return; |
| } |
| ReportSuccess(effective_test_name); |
| } |
| |
| void ConformanceTestSuite::ExpectParseFailureForJson( |
| const string& test_name, ConformanceLevel level, const string& input_json) { |
| ConformanceRequest request; |
| ConformanceResponse response; |
| request.set_json_payload(input_json); |
| string effective_test_name = |
| ConformanceLevelToString(level) + ".JsonInput." + test_name; |
| |
| // We don't expect output, but if the program erroneously accepts the protobuf |
| // we let it send its response as this. We must not leave it unspecified. |
| request.set_requested_output_format(conformance::JSON); |
| |
| RunTest(effective_test_name, request, &response); |
| if (response.result_case() == ConformanceResponse::kParseError) { |
| ReportSuccess(effective_test_name); |
| } else if (response.result_case() == ConformanceResponse::kSkipped) { |
| ReportSkip(effective_test_name, request, response); |
| } else { |
| ReportFailure(effective_test_name, level, request, response, |
| "Should have failed to parse, but didn't."); |
| } |
| } |
| |
| void ConformanceTestSuite::ExpectSerializeFailureForJson( |
| const string& test_name, ConformanceLevel level, const string& text_format) { |
| TestAllTypes payload_message; |
| GOOGLE_CHECK( |
| TextFormat::ParseFromString(text_format, &payload_message)) |
| << "Failed to parse: " << text_format; |
| |
| ConformanceRequest request; |
| ConformanceResponse response; |
| request.set_protobuf_payload(payload_message.SerializeAsString()); |
| string effective_test_name = |
| ConformanceLevelToString(level) + "." + test_name + ".JsonOutput"; |
| request.set_requested_output_format(conformance::JSON); |
| |
| RunTest(effective_test_name, request, &response); |
| if (response.result_case() == ConformanceResponse::kSerializeError) { |
| ReportSuccess(effective_test_name); |
| } else if (response.result_case() == ConformanceResponse::kSkipped) { |
| ReportSkip(effective_test_name, request, response); |
| } else { |
| ReportFailure(effective_test_name, level, request, response, |
| "Should have failed to serialize, but didn't."); |
| } |
| } |
| |
| void ConformanceTestSuite::TestPrematureEOFForType(FieldDescriptor::Type type) { |
| // Incomplete values for each wire type. |
| static const string incompletes[6] = { |
| string("\x80"), // VARINT |
| string("abcdefg"), // 64BIT |
| string("\x80"), // DELIMITED (partial length) |
| string(), // START_GROUP (no value required) |
| string(), // END_GROUP (no value required) |
| string("abc") // 32BIT |
| }; |
| |
| const FieldDescriptor* field = GetFieldForType(type, false); |
| const FieldDescriptor* rep_field = GetFieldForType(type, true); |
| WireFormatLite::WireType wire_type = WireFormatLite::WireTypeForFieldType( |
| static_cast<WireFormatLite::FieldType>(type)); |
| const string& incomplete = incompletes[wire_type]; |
| const string type_name = |
| UpperCase(string(".") + FieldDescriptor::TypeName(type)); |
| |
| ExpectParseFailureForProto( |
| tag(field->number(), wire_type), |
| "PrematureEofBeforeKnownNonRepeatedValue" + type_name, REQUIRED); |
| |
| ExpectParseFailureForProto( |
| tag(rep_field->number(), wire_type), |
| "PrematureEofBeforeKnownRepeatedValue" + type_name, REQUIRED); |
| |
| ExpectParseFailureForProto( |
| tag(UNKNOWN_FIELD, wire_type), |
| "PrematureEofBeforeUnknownValue" + type_name, REQUIRED); |
| |
| ExpectParseFailureForProto( |
| cat( tag(field->number(), wire_type), incomplete ), |
| "PrematureEofInsideKnownNonRepeatedValue" + type_name, REQUIRED); |
| |
| ExpectParseFailureForProto( |
| cat( tag(rep_field->number(), wire_type), incomplete ), |
| "PrematureEofInsideKnownRepeatedValue" + type_name, REQUIRED); |
| |
| ExpectParseFailureForProto( |
| cat( tag(UNKNOWN_FIELD, wire_type), incomplete ), |
| "PrematureEofInsideUnknownValue" + type_name, REQUIRED); |
| |
| if (wire_type == WireFormatLite::WIRETYPE_LENGTH_DELIMITED) { |
| ExpectParseFailureForProto( |
| cat( tag(field->number(), wire_type), varint(1) ), |
| "PrematureEofInDelimitedDataForKnownNonRepeatedValue" + type_name, |
| REQUIRED); |
| |
| ExpectParseFailureForProto( |
| cat( tag(rep_field->number(), wire_type), varint(1) ), |
| "PrematureEofInDelimitedDataForKnownRepeatedValue" + type_name, |
| REQUIRED); |
| |
| // EOF in the middle of delimited data for unknown value. |
| ExpectParseFailureForProto( |
| cat( tag(UNKNOWN_FIELD, wire_type), varint(1) ), |
| "PrematureEofInDelimitedDataForUnknownValue" + type_name, REQUIRED); |
| |
| if (type == FieldDescriptor::TYPE_MESSAGE) { |
| // Submessage ends in the middle of a value. |
| string incomplete_submsg = |
| cat( tag(WireFormatLite::TYPE_INT32, WireFormatLite::WIRETYPE_VARINT), |
| incompletes[WireFormatLite::WIRETYPE_VARINT] ); |
| ExpectHardParseFailureForProto( |
| cat( tag(field->number(), WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
| varint(incomplete_submsg.size()), |
| incomplete_submsg ), |
| "PrematureEofInSubmessageValue" + type_name, REQUIRED); |
| } |
| } else if (type != FieldDescriptor::TYPE_GROUP) { |
| // Non-delimited, non-group: eligible for packing. |
| |
| // Packed region ends in the middle of a value. |
| ExpectHardParseFailureForProto( |
| cat(tag(rep_field->number(), WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
| varint(incomplete.size()), incomplete), |
| "PrematureEofInPackedFieldValue" + type_name, REQUIRED); |
| |
| // EOF in the middle of packed region. |
| ExpectParseFailureForProto( |
| cat(tag(rep_field->number(), WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
| varint(1)), |
| "PrematureEofInPackedField" + type_name, REQUIRED); |
| } |
| } |
| |
| void ConformanceTestSuite::TestValidDataForType( |
| FieldDescriptor::Type type, |
| std::vector<std::pair<std::string, std::string>> values) { |
| const string type_name = |
| UpperCase(string(".") + FieldDescriptor::TypeName(type)); |
| WireFormatLite::WireType wire_type = WireFormatLite::WireTypeForFieldType( |
| static_cast<WireFormatLite::FieldType>(type)); |
| const FieldDescriptor* field = GetFieldForType(type, false); |
| const FieldDescriptor* rep_field = GetFieldForType(type, true); |
| |
| RunValidProtobufTest("ValidDataScalar" + type_name, REQUIRED, |
| cat(tag(field->number(), wire_type), values[0].first), |
| field->name() + ": " + values[0].second); |
| |
| string proto; |
| string text = field->name() + ": " + values.back().second; |
| for (size_t i = 0; i < values.size(); i++) { |
| proto += cat(tag(field->number(), wire_type), values[i].first); |
| } |
| RunValidProtobufTest("RepeatedScalarSelectsLast" + type_name, REQUIRED, |
| proto, text); |
| |
| proto.clear(); |
| text.clear(); |
| |
| for (size_t i = 0; i < values.size(); i++) { |
| proto += cat(tag(rep_field->number(), wire_type), values[i].first); |
| text += rep_field->name() + ": " + values[i].second + " "; |
| } |
| RunValidProtobufTest("ValidDataRepeated" + type_name, REQUIRED, proto, text); |
| } |
| |
| void ConformanceTestSuite::SetFailureList(const string& filename, |
| const vector<string>& failure_list) { |
| failure_list_filename_ = filename; |
| expected_to_fail_.clear(); |
| std::copy(failure_list.begin(), failure_list.end(), |
| std::inserter(expected_to_fail_, expected_to_fail_.end())); |
| } |
| |
| bool ConformanceTestSuite::CheckSetEmpty(const std::set<string>& set_to_check, |
| const std::string& write_to_file, |
| const std::string& msg) { |
| if (set_to_check.empty()) { |
| return true; |
| } else { |
| StringAppendF(&output_, "\n"); |
| StringAppendF(&output_, "%s\n\n", msg.c_str()); |
| for (std::set<string>::const_iterator iter = set_to_check.begin(); |
| iter != set_to_check.end(); ++iter) { |
| StringAppendF(&output_, " %s\n", iter->c_str()); |
| } |
| StringAppendF(&output_, "\n"); |
| |
| if (!write_to_file.empty()) { |
| std::ofstream os(write_to_file); |
| if (os) { |
| for (std::set<string>::const_iterator iter = set_to_check.begin(); |
| iter != set_to_check.end(); ++iter) { |
| os << *iter << "\n"; |
| } |
| } else { |
| StringAppendF(&output_, "Failed to open file: %s\n", |
| write_to_file.c_str()); |
| } |
| } |
| |
| return false; |
| } |
| } |
| |
| void ConformanceTestSuite::TestIllegalTags() { |
| // field num 0 is illegal |
| string nullfield[] = { |
| "\1DEADBEEF", |
| "\2\1\1", |
| "\3\4", |
| "\5DEAD" |
| }; |
| for (int i = 0; i < 4; i++) { |
| string name = "IllegalZeroFieldNum_Case_0"; |
| name.back() += i; |
| ExpectParseFailureForProto(nullfield[i], name, REQUIRED); |
| } |
| } |
| |
| bool ConformanceTestSuite::RunSuite(ConformanceTestRunner* runner, |
| std::string* output) { |
| runner_ = runner; |
| successes_ = 0; |
| expected_failures_ = 0; |
| skipped_.clear(); |
| test_names_.clear(); |
| unexpected_failing_tests_.clear(); |
| unexpected_succeeding_tests_.clear(); |
| type_resolver_.reset(NewTypeResolverForDescriptorPool( |
| kTypeUrlPrefix, DescriptorPool::generated_pool())); |
| type_url_ = GetTypeUrl(TestAllTypes::descriptor()); |
| |
| output_ = "\nCONFORMANCE TEST BEGIN ====================================\n\n"; |
| |
| for (int i = 1; i <= FieldDescriptor::MAX_TYPE; i++) { |
| if (i == FieldDescriptor::TYPE_GROUP) continue; |
| TestPrematureEOFForType(static_cast<FieldDescriptor::Type>(i)); |
| } |
| |
| TestIllegalTags(); |
| |
| int64 kInt64Min = -9223372036854775808ULL; |
| int64 kInt64Max = 9223372036854775807ULL; |
| uint64 kUint64Max = 18446744073709551615ULL; |
| int32 kInt32Max = 2147483647; |
| int32 kInt32Min = -2147483648; |
| uint32 kUint32Max = 4294967295UL; |
| |
| TestValidDataForType(FieldDescriptor::TYPE_DOUBLE, { |
| {dbl(0.1), "0.1"}, |
| {dbl(1.7976931348623157e+308), "1.7976931348623157e+308"}, |
| {dbl(2.22507385850720138309e-308), "2.22507385850720138309e-308"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_FLOAT, { |
| {flt(0.1), "0.1"}, |
| {flt(1.00000075e-36), "1.00000075e-36"}, |
| {flt(3.402823e+38), "3.402823e+38"}, // 3.40282347e+38 |
| {flt(1.17549435e-38f), "1.17549435e-38"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_INT64, { |
| {varint(12345), "12345"}, |
| {varint(kInt64Max), std::to_string(kInt64Max)}, |
| {varint(kInt64Min), std::to_string(kInt64Min)} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_UINT64, { |
| {varint(12345), "12345"}, |
| {varint(kUint64Max), std::to_string(kUint64Max)}, |
| {varint(0), "0"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_INT32, { |
| {varint(12345), "12345"}, |
| {longvarint(12345, 2), "12345"}, |
| {longvarint(12345, 7), "12345"}, |
| {varint(kInt32Max), std::to_string(kInt32Max)}, |
| {varint(kInt32Min), std::to_string(kInt32Min)}, |
| {varint(1LL << 33), std::to_string(static_cast<int32>(1LL << 33))}, |
| {varint((1LL << 33) - 1), |
| std::to_string(static_cast<int32>((1LL << 33) - 1))}, |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_UINT32, { |
| {varint(12345), "12345"}, |
| {longvarint(12345, 2), "12345"}, |
| {longvarint(12345, 7), "12345"}, |
| {varint(kUint32Max), std::to_string(kUint32Max)}, // UINT32_MAX |
| {varint(0), "0"}, |
| {varint(1LL << 33), std::to_string(static_cast<uint32>(1LL << 33))}, |
| {varint((1LL << 33) - 1), |
| std::to_string(static_cast<uint32>((1LL << 33) - 1))}, |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_FIXED64, { |
| {u64(12345), "12345"}, |
| {u64(kUint64Max), std::to_string(kUint64Max)}, |
| {u64(0), "0"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_FIXED32, { |
| {u32(12345), "12345"}, |
| {u32(kUint32Max), std::to_string(kUint32Max)}, // UINT32_MAX |
| {u32(0), "0"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_SFIXED64, { |
| {u64(12345), "12345"}, |
| {u64(kInt64Max), std::to_string(kInt64Max)}, |
| {u64(kInt64Min), std::to_string(kInt64Min)} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_SFIXED32, { |
| {u32(12345), "12345"}, |
| {u32(kInt32Max), std::to_string(kInt32Max)}, |
| {u32(kInt32Min), std::to_string(kInt32Min)} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_BOOL, { |
| {varint(1), "true"}, |
| {varint(0), "false"}, |
| {varint(12345678), "true"} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_SINT32, { |
| {zz32(12345), "12345"}, |
| {zz32(kInt32Max), std::to_string(kInt32Max)}, |
| {zz32(kInt32Min), std::to_string(kInt32Min)} |
| }); |
| TestValidDataForType(FieldDescriptor::TYPE_SINT64, { |
| {zz64(12345), "12345"}, |
| {zz64(kInt64Max), std::to_string(kInt64Max)}, |
| {zz64(kInt64Min), std::to_string(kInt64Min)} |
| }); |
| |
| // TODO(haberman): |
| // TestValidDataForType(FieldDescriptor::TYPE_STRING |
| // TestValidDataForType(FieldDescriptor::TYPE_GROUP |
| // TestValidDataForType(FieldDescriptor::TYPE_MESSAGE |
| // TestValidDataForType(FieldDescriptor::TYPE_BYTES |
| // TestValidDataForType(FieldDescriptor::TYPE_ENUM |
| |
| RunValidJsonTest("HelloWorld", REQUIRED, |
| "{\"optionalString\":\"Hello, World!\"}", |
| "optional_string: 'Hello, World!'"); |
| |
| // NOTE: The spec for JSON support is still being sorted out, these may not |
| // all be correct. |
| // Test field name conventions. |
| RunValidJsonTest( |
| "FieldNameInSnakeCase", REQUIRED, |
| R"({ |
| "fieldname1": 1, |
| "fieldName2": 2, |
| "FieldName3": 3, |
| "fieldName4": 4 |
| })", |
| R"( |
| fieldname1: 1 |
| field_name2: 2 |
| _field_name3: 3 |
| field__name4_: 4 |
| )"); |
| RunValidJsonTest( |
| "FieldNameWithNumbers", REQUIRED, |
| R"({ |
| "field0name5": 5, |
| "field0Name6": 6 |
| })", |
| R"( |
| field0name5: 5 |
| field_0_name6: 6 |
| )"); |
| RunValidJsonTest( |
| "FieldNameWithMixedCases", REQUIRED, |
| R"({ |
| "fieldName7": 7, |
| "FieldName8": 8, |
| "fieldName9": 9, |
| "FieldName10": 10, |
| "FIELDNAME11": 11, |
| "FIELDName12": 12 |
| })", |
| R"( |
| fieldName7: 7 |
| FieldName8: 8 |
| field_Name9: 9 |
| Field_Name10: 10 |
| FIELD_NAME11: 11 |
| FIELD_name12: 12 |
| )"); |
| RunValidJsonTest( |
| "FieldNameWithDoubleUnderscores", RECOMMENDED, |
| R"({ |
| "FieldName13": 13, |
| "FieldName14": 14, |
| "fieldName15": 15, |
| "fieldName16": 16, |
| "fieldName17": 17, |
| "FieldName18": 18 |
| })", |
| R"( |
| __field_name13: 13 |
| __Field_name14: 14 |
| field__name15: 15 |
| field__Name16: 16 |
| field_name17__: 17 |
| Field_name18__: 18 |
| )"); |
| // Using the original proto field name in JSON is also allowed. |
| RunValidJsonTest( |
| "OriginalProtoFieldName", REQUIRED, |
| R"({ |
| "fieldname1": 1, |
| "field_name2": 2, |
| "_field_name3": 3, |
| "field__name4_": 4, |
| "field0name5": 5, |
| "field_0_name6": 6, |
| "fieldName7": 7, |
| "FieldName8": 8, |
| "field_Name9": 9, |
| "Field_Name10": 10, |
| "FIELD_NAME11": 11, |
| "FIELD_name12": 12, |
| "__field_name13": 13, |
| "__Field_name14": 14, |
| "field__name15": 15, |
| "field__Name16": 16, |
| "field_name17__": 17, |
| "Field_name18__": 18 |
| })", |
| R"( |
| fieldname1: 1 |
| field_name2: 2 |
| _field_name3: 3 |
| field__name4_: 4 |
| field0name5: 5 |
| field_0_name6: 6 |
| fieldName7: 7 |
| FieldName8: 8 |
| field_Name9: 9 |
| Field_Name10: 10 |
| FIELD_NAME11: 11 |
| FIELD_name12: 12 |
| __field_name13: 13 |
| __Field_name14: 14 |
| field__name15: 15 |
| field__Name16: 16 |
| field_name17__: 17 |
| Field_name18__: 18 |
| )"); |
| // Field names can be escaped. |
| RunValidJsonTest( |
| "FieldNameEscaped", REQUIRED, |
| R"({"fieldn\u0061me1": 1})", |
| "fieldname1: 1"); |
| // String ends with escape character. |
| ExpectParseFailureForJson( |
| "StringEndsWithEscapeChar", RECOMMENDED, |
| "{\"optionalString\": \"abc\\"); |
| // Field names must be quoted (or it's not valid JSON). |
| ExpectParseFailureForJson( |
| "FieldNameNotQuoted", RECOMMENDED, |
| "{fieldname1: 1}"); |
| // Trailing comma is not allowed (not valid JSON). |
| ExpectParseFailureForJson( |
| "TrailingCommaInAnObject", RECOMMENDED, |
| R"({"fieldname1":1,})"); |
| ExpectParseFailureForJson( |
| "TrailingCommaInAnObjectWithSpace", RECOMMENDED, |
| R"({"fieldname1":1 ,})"); |
| ExpectParseFailureForJson( |
| "TrailingCommaInAnObjectWithSpaceCommaSpace", RECOMMENDED, |
| R"({"fieldname1":1 , })"); |
| ExpectParseFailureForJson( |
| "TrailingCommaInAnObjectWithNewlines", RECOMMENDED, |
| R"({ |
| "fieldname1":1, |
| })"); |
| // JSON doesn't support comments. |
| ExpectParseFailureForJson( |
| "JsonWithComments", RECOMMENDED, |
| R"({ |
| // This is a comment. |
| "fieldname1": 1 |
| })"); |
| // JSON spec says whitespace doesn't matter, so try a few spacings to be sure. |
| RunValidJsonTest( |
| "OneLineNoSpaces", RECOMMENDED, |
| "{\"optionalInt32\":1,\"optionalInt64\":2}", |
| R"( |
| optional_int32: 1 |
| optional_int64: 2 |
| )"); |
| RunValidJsonTest( |
| "OneLineWithSpaces", RECOMMENDED, |
| "{ \"optionalInt32\" : 1 , \"optionalInt64\" : 2 }", |
| R"( |
| optional_int32: 1 |
| optional_int64: 2 |
| )"); |
| RunValidJsonTest( |
| "MultilineNoSpaces", RECOMMENDED, |
| "{\n\"optionalInt32\"\n:\n1\n,\n\"optionalInt64\"\n:\n2\n}", |
| R"( |
| optional_int32: 1 |
| optional_int64: 2 |
| )"); |
| RunValidJsonTest( |
| "MultilineWithSpaces", RECOMMENDED, |
| "{\n \"optionalInt32\" : 1\n ,\n \"optionalInt64\" : 2\n}\n", |
| R"( |
| optional_int32: 1 |
| optional_int64: 2 |
| )"); |
| // Missing comma between key/value pairs. |
| ExpectParseFailureForJson( |
| "MissingCommaOneLine", RECOMMENDED, |
| "{ \"optionalInt32\": 1 \"optionalInt64\": 2 }"); |
| ExpectParseFailureForJson( |
| "MissingCommaMultiline", RECOMMENDED, |
| "{\n \"optionalInt32\": 1\n \"optionalInt64\": 2\n}"); |
| // Duplicated field names are not allowed. |
| ExpectParseFailureForJson( |
| "FieldNameDuplicate", RECOMMENDED, |
| R"({ |
| "optionalNestedMessage": {a: 1}, |
| "optionalNestedMessage": {} |
| })"); |
| ExpectParseFailureForJson( |
| "FieldNameDuplicateDifferentCasing1", RECOMMENDED, |
| R"({ |
| "optional_nested_message": {a: 1}, |
| "optionalNestedMessage": {} |
| })"); |
| ExpectParseFailureForJson( |
| "FieldNameDuplicateDifferentCasing2", RECOMMENDED, |
| R"({ |
| "optionalNestedMessage": {a: 1}, |
| "optional_nested_message": {} |
| })"); |
| // Serializers should use lowerCamelCase by default. |
| RunValidJsonTestWithValidator( |
| "FieldNameInLowerCamelCase", REQUIRED, |
| R"({ |
| "fieldname1": 1, |
| "fieldName2": 2, |
| "FieldName3": 3, |
| "fieldName4": 4 |
| })", |
| [](const Json::Value& value) { |
| return value.isMember("fieldname1") && |
| value.isMember("fieldName2") && |
| value.isMember("FieldName3") && |
| value.isMember("fieldName4"); |
| }); |
| RunValidJsonTestWithValidator( |
| "FieldNameWithNumbers", REQUIRED, |
| R"({ |
| "field0name5": 5, |
| "field0Name6": 6 |
| })", |
| [](const Json::Value& value) { |
| return value.isMember("field0name5") && |
| value.isMember("field0Name6"); |
| }); |
| RunValidJsonTestWithValidator( |
| "FieldNameWithMixedCases", REQUIRED, |
| R"({ |
| "fieldName7": 7, |
| "FieldName8": 8, |
| "fieldName9": 9, |
| "FieldName10": 10, |
| "FIELDNAME11": 11, |
| "FIELDName12": 12 |
| })", |
| [](const Json::Value& value) { |
| return value.isMember("fieldName7") && |
| value.isMember("FieldName8") && |
| value.isMember("fieldName9") && |
| value.isMember("FieldName10") && |
| value.isMember("FIELDNAME11") && |
| value.isMember("FIELDName12"); |
| }); |
| RunValidJsonTestWithValidator( |
| "FieldNameWithDoubleUnderscores", RECOMMENDED, |
| R"({ |
| "FieldName13": 13, |
| "FieldName14": 14, |
| "fieldName15": 15, |
| "fieldName16": 16, |
| "fieldName17": 17, |
| "FieldName18": 18 |
| })", |
| [](const Json::Value& value) { |
| return value.isMember("FieldName13") && |
| value.isMember("FieldName14") && |
| value.isMember("fieldName15") && |
| value.isMember("fieldName16") && |
| value.isMember("fieldName17") && |
| value.isMember("FieldName18"); |
| }); |
| |
| // Integer fields. |
| RunValidJsonTest( |
| "Int32FieldMaxValue", REQUIRED, |
| R"({"optionalInt32": 2147483647})", |
| "optional_int32: 2147483647"); |
| RunValidJsonTest( |
| "Int32FieldMinValue", REQUIRED, |
| R"({"optionalInt32": -2147483648})", |
| "optional_int32: -2147483648"); |
| RunValidJsonTest( |
| "Uint32FieldMaxValue", REQUIRED, |
| R"({"optionalUint32": 4294967295})", |
| "optional_uint32: 4294967295"); |
| RunValidJsonTest( |
| "Int64FieldMaxValue", REQUIRED, |
| R"({"optionalInt64": "9223372036854775807"})", |
| "optional_int64: 9223372036854775807"); |
| RunValidJsonTest( |
| "Int64FieldMinValue", REQUIRED, |
| R"({"optionalInt64": "-9223372036854775808"})", |
| "optional_int64: -9223372036854775808"); |
| RunValidJsonTest( |
| "Uint64FieldMaxValue", REQUIRED, |
| R"({"optionalUint64": "18446744073709551615"})", |
| "optional_uint64: 18446744073709551615"); |
| // While not the largest Int64, this is the largest |
| // Int64 which can be exactly represented within an |
| // IEEE-754 64-bit float, which is the expected level |
| // of interoperability guarantee. Larger values may |
| // work in some implementations, but should not be |
| // relied upon. |
| RunValidJsonTest( |
| "Int64FieldMaxValueNotQuoted", REQUIRED, |
| R"({"optionalInt64": 9223372036854774784})", |
| "optional_int64: 9223372036854774784"); |
| RunValidJsonTest( |
| "Int64FieldMinValueNotQuoted", REQUIRED, |
| R"({"optionalInt64": -9223372036854775808})", |
| "optional_int64: -9223372036854775808"); |
| // Largest interoperable Uint64; see comment above |
| // for Int64FieldMaxValueNotQuoted. |
| RunValidJsonTest( |
| "Uint64FieldMaxValueNotQuoted", REQUIRED, |
| R"({"optionalUint64": 18446744073709549568})", |
| "optional_uint64: 18446744073709549568"); |
| // Values can be represented as JSON strings. |
| RunValidJsonTest( |
| "Int32FieldStringValue", REQUIRED, |
| R"({"optionalInt32": "2147483647"})", |
| "optional_int32: 2147483647"); |
| RunValidJsonTest( |
| "Int32FieldStringValueEscaped", REQUIRED, |
| R"({"optionalInt32": "2\u003147483647"})", |
| "optional_int32: 2147483647"); |
| |
| // Parsers reject out-of-bound integer values. |
| ExpectParseFailureForJson( |
| "Int32FieldTooLarge", REQUIRED, |
| R"({"optionalInt32": 2147483648})"); |
| ExpectParseFailureForJson( |
| "Int32FieldTooSmall", REQUIRED, |
| R"({"optionalInt32": -2147483649})"); |
| ExpectParseFailureForJson( |
| "Uint32FieldTooLarge", REQUIRED, |
| R"({"optionalUint32": 4294967296})"); |
| ExpectParseFailureForJson( |
| "Int64FieldTooLarge", REQUIRED, |
| R"({"optionalInt64": "9223372036854775808"})"); |
| ExpectParseFailureForJson( |
| "Int64FieldTooSmall", REQUIRED, |
| R"({"optionalInt64": "-9223372036854775809"})"); |
| ExpectParseFailureForJson( |
| "Uint64FieldTooLarge", REQUIRED, |
| R"({"optionalUint64": "18446744073709551616"})"); |
| // Parser reject non-integer numeric values as well. |
| ExpectParseFailureForJson( |
| "Int32FieldNotInteger", REQUIRED, |
| R"({"optionalInt32": 0.5})"); |
| ExpectParseFailureForJson( |
| "Uint32FieldNotInteger", REQUIRED, |
| R"({"optionalUint32": 0.5})"); |
| ExpectParseFailureForJson( |
| "Int64FieldNotInteger", REQUIRED, |
| R"({"optionalInt64": "0.5"})"); |
| ExpectParseFailureForJson( |
| "Uint64FieldNotInteger", REQUIRED, |
| R"({"optionalUint64": "0.5"})"); |
| |
| // Integers but represented as float values are accepted. |
| RunValidJsonTest( |
| "Int32FieldFloatTrailingZero", REQUIRED, |
| R"({"optionalInt32": 100000.000})", |
| "optional_int32: 100000"); |
| RunValidJsonTest( |
| "Int32FieldExponentialFormat", REQUIRED, |
| R"({"optionalInt32": 1e5})", |
| "optional_int32: 100000"); |
| RunValidJsonTest( |
| "Int32FieldMaxFloatValue", REQUIRED, |
| R"({"optionalInt32": 2.147483647e9})", |
| "optional_int32: 2147483647"); |
| RunValidJsonTest( |
| "Int32FieldMinFloatValue", REQUIRED, |
| R"({"optionalInt32": -2.147483648e9})", |
| "optional_int32: -2147483648"); |
| RunValidJsonTest( |
| "Uint32FieldMaxFloatValue", REQUIRED, |
| R"({"optionalUint32": 4.294967295e9})", |
| "optional_uint32: 4294967295"); |
| |
| // Parser reject non-numeric values. |
| ExpectParseFailureForJson( |
| "Int32FieldNotNumber", REQUIRED, |
| R"({"optionalInt32": "3x3"})"); |
| ExpectParseFailureForJson( |
| "Uint32FieldNotNumber", REQUIRED, |
| R"({"optionalUint32": "3x3"})"); |
| ExpectParseFailureForJson( |
| "Int64FieldNotNumber", REQUIRED, |
| R"({"optionalInt64": "3x3"})"); |
| ExpectParseFailureForJson( |
| "Uint64FieldNotNumber", REQUIRED, |
| R"({"optionalUint64": "3x3"})"); |
| // JSON does not allow "+" on numric values. |
| ExpectParseFailureForJson( |
| "Int32FieldPlusSign", REQUIRED, |
| R"({"optionalInt32": +1})"); |
| // JSON doesn't allow leading 0s. |
| ExpectParseFailureForJson( |
| "Int32FieldLeadingZero", REQUIRED, |
| R"({"optionalInt32": 01})"); |
| ExpectParseFailureForJson( |
| "Int32FieldNegativeWithLeadingZero", REQUIRED, |
| R"({"optionalInt32": -01})"); |
| // String values must follow the same syntax rule. Specifically leading |
| // or trailing spaces are not allowed. |
| ExpectParseFailureForJson( |
| "Int32FieldLeadingSpace", REQUIRED, |
| R"({"optionalInt32": " 1"})"); |
| ExpectParseFailureForJson( |
| "Int32FieldTrailingSpace", REQUIRED, |
| R"({"optionalInt32": "1 "})"); |
| |
| // 64-bit values are serialized as strings. |
| RunValidJsonTestWithValidator( |
| "Int64FieldBeString", RECOMMENDED, |
| R"({"optionalInt64": 1})", |
| [](const Json::Value& value) { |
| return value["optionalInt64"].type() == Json::stringValue && |
| value["optionalInt64"].asString() == "1"; |
| }); |
| RunValidJsonTestWithValidator( |
| "Uint64FieldBeString", RECOMMENDED, |
| R"({"optionalUint64": 1})", |
| [](const Json::Value& value) { |
| return value["optionalUint64"].type() == Json::stringValue && |
| value["optionalUint64"].asString() == "1"; |
| }); |
| |
| // Bool fields. |
| RunValidJsonTest( |
| "BoolFieldTrue", REQUIRED, |
| R"({"optionalBool":true})", |
| "optional_bool: true"); |
| RunValidJsonTest( |
| "BoolFieldFalse", REQUIRED, |
| R"({"optionalBool":false})", |
| "optional_bool: false"); |
| |
| // Other forms are not allowed. |
| ExpectParseFailureForJson( |
| "BoolFieldIntegerZero", RECOMMENDED, |
| R"({"optionalBool":0})"); |
| ExpectParseFailureForJson( |
| "BoolFieldIntegerOne", RECOMMENDED, |
| R"({"optionalBool":1})"); |
| ExpectParseFailureForJson( |
| "BoolFieldCamelCaseTrue", RECOMMENDED, |
| R"({"optionalBool":True})"); |
| ExpectParseFailureForJson( |
| "BoolFieldCamelCaseFalse", RECOMMENDED, |
| R"({"optionalBool":False})"); |
| ExpectParseFailureForJson( |
| "BoolFieldAllCapitalTrue", RECOMMENDED, |
| R"({"optionalBool":TRUE})"); |
| ExpectParseFailureForJson( |
| "BoolFieldAllCapitalFalse", RECOMMENDED, |
| R"({"optionalBool":FALSE})"); |
| ExpectParseFailureForJson( |
| "BoolFieldDoubleQuotedTrue", RECOMMENDED, |
| R"({"optionalBool":"true"})"); |
| ExpectParseFailureForJson( |
| "BoolFieldDoubleQuotedFalse", RECOMMENDED, |
| R"({"optionalBool":"false"})"); |
| |
| // Float fields. |
| RunValidJsonTest( |
| "FloatFieldMinPositiveValue", REQUIRED, |
| R"({"optionalFloat": 1.175494e-38})", |
| "optional_float: 1.175494e-38"); |
| RunValidJsonTest( |
| "FloatFieldMaxNegativeValue", REQUIRED, |
| R"({"optionalFloat": -1.175494e-38})", |
| "optional_float: -1.175494e-38"); |
| RunValidJsonTest( |
| "FloatFieldMaxPositiveValue", REQUIRED, |
| R"({"optionalFloat": 3.402823e+38})", |
| "optional_float: 3.402823e+38"); |
| RunValidJsonTest( |
| "FloatFieldMinNegativeValue", REQUIRED, |
| R"({"optionalFloat": 3.402823e+38})", |
| "optional_float: 3.402823e+38"); |
| // Values can be quoted. |
| RunValidJsonTest( |
| "FloatFieldQuotedValue", REQUIRED, |
| R"({"optionalFloat": "1"})", |
| "optional_float: 1"); |
| // Special values. |
| RunValidJsonTest( |
| "FloatFieldNan", REQUIRED, |
| R"({"optionalFloat": "NaN"})", |
| "optional_float: nan"); |
| RunValidJsonTest( |
| "FloatFieldInfinity", REQUIRED, |
| R"({"optionalFloat": "Infinity"})", |
| "optional_float: inf"); |
| RunValidJsonTest( |
| "FloatFieldNegativeInfinity", REQUIRED, |
| R"({"optionalFloat": "-Infinity"})", |
| "optional_float: -inf"); |
| // Non-cannonical Nan will be correctly normalized. |
| { |
| TestAllTypes message; |
| // IEEE floating-point standard 32-bit quiet NaN: |
| // 0111 1111 1xxx xxxx xxxx xxxx xxxx xxxx |
| message.set_optional_float( |
| WireFormatLite::DecodeFloat(0x7FA12345)); |
| RunValidJsonTestWithProtobufInput( |
| "FloatFieldNormalizeQuietNan", REQUIRED, message, |
| "optional_float: nan"); |
| // IEEE floating-point standard 64-bit signaling NaN: |
| // 1111 1111 1xxx xxxx xxxx xxxx xxxx xxxx |
| message.set_optional_float( |
| WireFormatLite::DecodeFloat(0xFFB54321)); |
| RunValidJsonTestWithProtobufInput( |
| "FloatFieldNormalizeSignalingNan", REQUIRED, message, |
| "optional_float: nan"); |
| } |
| |
| // Special values must be quoted. |
| ExpectParseFailureForJson( |
| "FloatFieldNanNotQuoted", RECOMMENDED, |
| R"({"optionalFloat": NaN})"); |
| ExpectParseFailureForJson( |
| "FloatFieldInfinityNotQuoted", RECOMMENDED, |
| R"({"optionalFloat": Infinity})"); |
| ExpectParseFailureForJson( |
| "FloatFieldNegativeInfinityNotQuoted", RECOMMENDED, |
| R"({"optionalFloat": -Infinity})"); |
| // Parsers should reject out-of-bound values. |
| ExpectParseFailureForJson( |
| "FloatFieldTooSmall", REQUIRED, |
| R"({"optionalFloat": -3.502823e+38})"); |
| ExpectParseFailureForJson( |
| "FloatFieldTooLarge", REQUIRED, |
| R"({"optionalFloat": 3.502823e+38})"); |
| |
| // Double fields. |
| RunValidJsonTest( |
| "DoubleFieldMinPositiveValue", REQUIRED, |
| R"({"optionalDouble": 2.22507e-308})", |
| "optional_double: 2.22507e-308"); |
| RunValidJsonTest( |
| "DoubleFieldMaxNegativeValue", REQUIRED, |
| R"({"optionalDouble": -2.22507e-308})", |
| "optional_double: -2.22507e-308"); |
| RunValidJsonTest( |
| "DoubleFieldMaxPositiveValue", REQUIRED, |
| R"({"optionalDouble": 1.79769e+308})", |
| "optional_double: 1.79769e+308"); |
| RunValidJsonTest( |
| "DoubleFieldMinNegativeValue", REQUIRED, |
| R"({"optionalDouble": -1.79769e+308})", |
| "optional_double: -1.79769e+308"); |
| // Values can be quoted. |
| RunValidJsonTest( |
| "DoubleFieldQuotedValue", REQUIRED, |
| R"({"optionalDouble": "1"})", |
| "optional_double: 1"); |
| // Speical values. |
| RunValidJsonTest( |
| "DoubleFieldNan", REQUIRED, |
| R"({"optionalDouble": "NaN"})", |
| "optional_double: nan"); |
| RunValidJsonTest( |
| "DoubleFieldInfinity", REQUIRED, |
| R"({"optionalDouble": "Infinity"})", |
| "optional_double: inf"); |
| RunValidJsonTest( |
| "DoubleFieldNegativeInfinity", REQUIRED, |
| R"({"optionalDouble": "-Infinity"})", |
| "optional_double: -inf"); |
| // Non-cannonical Nan will be correctly normalized. |
| { |
| TestAllTypes message; |
| message.set_optional_double( |
| WireFormatLite::DecodeDouble(0x7FFA123456789ABCLL)); |
| RunValidJsonTestWithProtobufInput( |
| "DoubleFieldNormalizeQuietNan", REQUIRED, message, |
| "optional_double: nan"); |
| message.set_optional_double( |
| WireFormatLite::DecodeDouble(0xFFFBCBA987654321LL)); |
| RunValidJsonTestWithProtobufInput( |
| "DoubleFieldNormalizeSignalingNan", REQUIRED, message, |
| "optional_double: nan"); |
| } |
| |
| // Special values must be quoted. |
| ExpectParseFailureForJson( |
| "DoubleFieldNanNotQuoted", RECOMMENDED, |
| R"({"optionalDouble": NaN})"); |
| ExpectParseFailureForJson( |
| "DoubleFieldInfinityNotQuoted", RECOMMENDED, |
| R"({"optionalDouble": Infinity})"); |
| ExpectParseFailureForJson( |
| "DoubleFieldNegativeInfinityNotQuoted", RECOMMENDED, |
| R"({"optionalDouble": -Infinity})"); |
| |
| // Parsers should reject out-of-bound values. |
| ExpectParseFailureForJson( |
| "DoubleFieldTooSmall", REQUIRED, |
| R"({"optionalDouble": -1.89769e+308})"); |
| ExpectParseFailureForJson( |
| "DoubleFieldTooLarge", REQUIRED, |
| R"({"optionalDouble": +1.89769e+308})"); |
| |
| // Enum fields. |
| RunValidJsonTest( |
| "EnumField", REQUIRED, |
| R"({"optionalNestedEnum": "FOO"})", |
| "optional_nested_enum: FOO"); |
| // Enum values must be represented as strings. |
| ExpectParseFailureForJson( |
| "EnumFieldNotQuoted", REQUIRED, |
| R"({"optionalNestedEnum": FOO})"); |
| // Numeric values are allowed. |
| RunValidJsonTest( |
| "EnumFieldNumericValueZero", REQUIRED, |
| R"({"optionalNestedEnum": 0})", |
| "optional_nested_enum: FOO"); |
| RunValidJsonTest( |
| "EnumFieldNumericValueNonZero", REQUIRED, |
| R"({"optionalNestedEnum": 1})", |
| "optional_nested_enum: BAR"); |
| // Unknown enum values are represented as numeric values. |
| RunValidJsonTestWithValidator( |
| "EnumFieldUnknownValue", REQUIRED, |
| R"({"optionalNestedEnum": 123})", |
| [](const Json::Value& value) { |
| return value["optionalNestedEnum"].type() == Json::intValue && |
| value["optionalNestedEnum"].asInt() == 123; |
| }); |
| |
| // String fields. |
| RunValidJsonTest( |
| "StringField", REQUIRED, |
| R"({"optionalString": "Hello world!"})", |
| "optional_string: \"Hello world!\""); |
| RunValidJsonTest( |
| "StringFieldUnicode", REQUIRED, |
| // Google in Chinese. |
| R"({"optionalString": "谷歌"})", |
| R"(optional_string: "谷歌")"); |
| RunValidJsonTest( |
| "StringFieldEscape", REQUIRED, |
| R"({"optionalString": "\"\\\/\b\f\n\r\t"})", |
| R"(optional_string: "\"\\/\b\f\n\r\t")"); |
| RunValidJsonTest( |
| "StringFieldUnicodeEscape", REQUIRED, |
| R"({"optionalString": "\u8C37\u6B4C"})", |
| R"(optional_string: "谷歌")"); |
| RunValidJsonTest( |
| "StringFieldUnicodeEscapeWithLowercaseHexLetters", REQUIRED, |
| R"({"optionalString": "\u8c37\u6b4c"})", |
| R"(optional_string: "谷歌")"); |
| RunValidJsonTest( |
| "StringFieldSurrogatePair", REQUIRED, |
| // The character is an emoji: grinning face with smiling eyes. 😁 |
| R"({"optionalString": "\uD83D\uDE01"})", |
| R"(optional_string: "\xF0\x9F\x98\x81")"); |
| |
| // Unicode escapes must start with "\u" (lowercase u). |
| ExpectParseFailureForJson( |
| "StringFieldUppercaseEscapeLetter", RECOMMENDED, |
| R"({"optionalString": "\U8C37\U6b4C"})"); |
| ExpectParseFailureForJson( |
| "StringFieldInvalidEscape", RECOMMENDED, |
| R"({"optionalString": "\uXXXX\u6B4C"})"); |
| ExpectParseFailureForJson( |
| "StringFieldUnterminatedEscape", RECOMMENDED, |
| R"({"optionalString": "\u8C3"})"); |
| ExpectParseFailureForJson( |
| "StringFieldUnpairedHighSurrogate", RECOMMENDED, |
| R"({"optionalString": "\uD800"})"); |
| ExpectParseFailureForJson( |
| "StringFieldUnpairedLowSurrogate", RECOMMENDED, |
| R"({"optionalString": "\uDC00"})"); |
| ExpectParseFailureForJson( |
| "StringFieldSurrogateInWrongOrder", RECOMMENDED, |
| R"({"optionalString": "\uDE01\uD83D"})"); |
| ExpectParseFailureForJson( |
| "StringFieldNotAString", REQUIRED, |
| R"({"optionalString": 12345})"); |
| |
| // Bytes fields. |
| RunValidJsonTest( |
| "BytesField", REQUIRED, |
| R"({"optionalBytes": "AQI="})", |
| R"(optional_bytes: "\x01\x02")"); |
| ExpectParseFailureForJson( |
| "BytesFieldInvalidBase64Characters", REQUIRED, |
| R"({"optionalBytes": "-_=="})"); |
| |
| // Message fields. |
| RunValidJsonTest( |
| "MessageField", REQUIRED, |
| R"({"optionalNestedMessage": {"a": 1234}})", |
| "optional_nested_message: {a: 1234}"); |
| |
| // Oneof fields. |
| ExpectParseFailureForJson( |
| "OneofFieldDuplicate", REQUIRED, |
| R"({"oneofUint32": 1, "oneofString": "test"})"); |
| // Ensure zero values for oneof make it out/backs. |
| { |
| TestAllTypes message; |
| message.set_oneof_uint32(0); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroUint32", RECOMMENDED, message, "oneof_uint32: 0"); |
| message.mutable_oneof_nested_message()->set_a(0); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroMessage", RECOMMENDED, message, "oneof_nested_message: {}"); |
| message.set_oneof_string(""); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroString", RECOMMENDED, message, "oneof_string: \"\""); |
| message.set_oneof_bytes(""); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroBytes", RECOMMENDED, message, "oneof_bytes: \"\""); |
| message.set_oneof_bool(false); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroBool", RECOMMENDED, message, "oneof_bool: false"); |
| message.set_oneof_uint64(0); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroUint64", RECOMMENDED, message, "oneof_uint64: 0"); |
| message.set_oneof_float(0.0f); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroFloat", RECOMMENDED, message, "oneof_float: 0"); |
| message.set_oneof_double(0.0); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroDouble", RECOMMENDED, message, "oneof_double: 0"); |
| message.set_oneof_enum(TestAllTypes::FOO); |
| RunValidProtobufTestWithMessage( |
| "OneofZeroEnum", RECOMMENDED, message, "oneof_enum: FOO"); |
| } |
| RunValidJsonTest( |
| "OneofZeroUint32", RECOMMENDED, |
| R"({"oneofUint32": 0})", "oneof_uint32: 0"); |
| RunValidJsonTest( |
| "OneofZeroMessage", RECOMMENDED, |
| R"({"oneofNestedMessage": {}})", "oneof_nested_message: {}"); |
| RunValidJsonTest( |
| "OneofZeroString", RECOMMENDED, |
| R"({"oneofString": ""})", "oneof_string: \"\""); |
| RunValidJsonTest( |
| "OneofZeroBytes", RECOMMENDED, |
| R"({"oneofBytes": ""})", "oneof_bytes: \"\""); |
| RunValidJsonTest( |
| "OneofZeroBool", RECOMMENDED, |
| R"({"oneofBool": false})", "oneof_bool: false"); |
| RunValidJsonTest( |
| "OneofZeroUint64", RECOMMENDED, |
| R"({"oneofUint64": 0})", "oneof_uint64: 0"); |
| RunValidJsonTest( |
| "OneofZeroFloat", RECOMMENDED, |
| R"({"oneofFloat": 0.0})", "oneof_float: 0"); |
| RunValidJsonTest( |
| "OneofZeroDouble", RECOMMENDED, |
| R"({"oneofDouble": 0.0})", "oneof_double: 0"); |
| RunValidJsonTest( |
| "OneofZeroEnum", RECOMMENDED, |
| R"({"oneofEnum":"FOO"})", "oneof_enum: FOO"); |
| |
| // Repeated fields. |
| RunValidJsonTest( |
| "PrimitiveRepeatedField", REQUIRED, |
| R"({"repeatedInt32": [1, 2, 3, 4]})", |
| "repeated_int32: [1, 2, 3, 4]"); |
| RunValidJsonTest( |
| "EnumRepeatedField", REQUIRED, |
| R"({"repeatedNestedEnum": ["FOO", "BAR", "BAZ"]})", |
| "repeated_nested_enum: [FOO, BAR, BAZ]"); |
| RunValidJsonTest( |
| "StringRepeatedField", REQUIRED, |
| R"({"repeatedString": ["Hello", "world"]})", |
| R"(repeated_string: ["Hello", "world"])"); |
| RunValidJsonTest( |
| "BytesRepeatedField", REQUIRED, |
| R"({"repeatedBytes": ["AAEC", "AQI="]})", |
| R"(repeated_bytes: ["\x00\x01\x02", "\x01\x02"])"); |
| RunValidJsonTest( |
| "MessageRepeatedField", REQUIRED, |
| R"({"repeatedNestedMessage": [{"a": 1234}, {"a": 5678}]})", |
| "repeated_nested_message: {a: 1234}" |
| "repeated_nested_message: {a: 5678}"); |
| |
| // Repeated field elements are of incorrect type. |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingIntegersGotBool", REQUIRED, |
| R"({"repeatedInt32": [1, false, 3, 4]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingIntegersGotString", REQUIRED, |
| R"({"repeatedInt32": [1, 2, "name", 4]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingIntegersGotMessage", REQUIRED, |
| R"({"repeatedInt32": [1, 2, 3, {"a": 4}]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingStringsGotInt", REQUIRED, |
| R"({"repeatedString": ["1", 2, "3", "4"]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingStringsGotBool", REQUIRED, |
| R"({"repeatedString": ["1", "2", false, "4"]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingStringsGotMessage", REQUIRED, |
| R"({"repeatedString": ["1", 2, "3", {"a": 4}]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingMessagesGotInt", REQUIRED, |
| R"({"repeatedNestedMessage": [{"a": 1}, 2]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingMessagesGotBool", REQUIRED, |
| R"({"repeatedNestedMessage": [{"a": 1}, false]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldWrongElementTypeExpectingMessagesGotString", REQUIRED, |
| R"({"repeatedNestedMessage": [{"a": 1}, "2"]})"); |
| // Trailing comma in the repeated field is not allowed. |
| ExpectParseFailureForJson( |
| "RepeatedFieldTrailingComma", RECOMMENDED, |
| R"({"repeatedInt32": [1, 2, 3, 4,]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldTrailingCommaWithSpace", RECOMMENDED, |
| "{\"repeatedInt32\": [1, 2, 3, 4 ,]}"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldTrailingCommaWithSpaceCommaSpace", RECOMMENDED, |
| "{\"repeatedInt32\": [1, 2, 3, 4 , ]}"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldTrailingCommaWithNewlines", RECOMMENDED, |
| "{\"repeatedInt32\": [\n 1,\n 2,\n 3,\n 4,\n]}"); |
| |
| // Map fields. |
| RunValidJsonTest( |
| "Int32MapField", REQUIRED, |
| R"({"mapInt32Int32": {"1": 2, "3": 4}})", |
| "map_int32_int32: {key: 1 value: 2}" |
| "map_int32_int32: {key: 3 value: 4}"); |
| ExpectParseFailureForJson( |
| "Int32MapFieldKeyNotQuoted", RECOMMENDED, |
| R"({"mapInt32Int32": {1: 2, 3: 4}})"); |
| RunValidJsonTest( |
| "Uint32MapField", REQUIRED, |
| R"({"mapUint32Uint32": {"1": 2, "3": 4}})", |
| "map_uint32_uint32: {key: 1 value: 2}" |
| "map_uint32_uint32: {key: 3 value: 4}"); |
| ExpectParseFailureForJson( |
| "Uint32MapFieldKeyNotQuoted", RECOMMENDED, |
| R"({"mapUint32Uint32": {1: 2, 3: 4}})"); |
| RunValidJsonTest( |
| "Int64MapField", REQUIRED, |
| R"({"mapInt64Int64": {"1": 2, "3": 4}})", |
| "map_int64_int64: {key: 1 value: 2}" |
| "map_int64_int64: {key: 3 value: 4}"); |
| ExpectParseFailureForJson( |
| "Int64MapFieldKeyNotQuoted", RECOMMENDED, |
| R"({"mapInt64Int64": {1: 2, 3: 4}})"); |
| RunValidJsonTest( |
| "Uint64MapField", REQUIRED, |
| R"({"mapUint64Uint64": {"1": 2, "3": 4}})", |
| "map_uint64_uint64: {key: 1 value: 2}" |
| "map_uint64_uint64: {key: 3 value: 4}"); |
| ExpectParseFailureForJson( |
| "Uint64MapFieldKeyNotQuoted", RECOMMENDED, |
| R"({"mapUint64Uint64": {1: 2, 3: 4}})"); |
| RunValidJsonTest( |
| "BoolMapField", REQUIRED, |
| R"({"mapBoolBool": {"true": true, "false": false}})", |
| "map_bool_bool: {key: true value: true}" |
| "map_bool_bool: {key: false value: false}"); |
| ExpectParseFailureForJson( |
| "BoolMapFieldKeyNotQuoted", RECOMMENDED, |
| R"({"mapBoolBool": {true: true, false: false}})"); |
| RunValidJsonTest( |
| "MessageMapField", REQUIRED, |
| R"({ |
| "mapStringNestedMessage": { |
| "hello": {"a": 1234}, |
| "world": {"a": 5678} |
| } |
| })", |
| R"( |
| map_string_nested_message: { |
| key: "hello" |
| value: {a: 1234} |
| } |
| map_string_nested_message: { |
| key: "world" |
| value: {a: 5678} |
| } |
| )"); |
| // Since Map keys are represented as JSON strings, escaping should be allowed. |
| RunValidJsonTest( |
| "Int32MapEscapedKey", REQUIRED, |
| R"({"mapInt32Int32": {"\u0031": 2}})", |
| "map_int32_int32: {key: 1 value: 2}"); |
| RunValidJsonTest( |
| "Int64MapEscapedKey", REQUIRED, |
| R"({"mapInt64Int64": {"\u0031": 2}})", |
| "map_int64_int64: {key: 1 value: 2}"); |
| RunValidJsonTest( |
| "BoolMapEscapedKey", REQUIRED, |
| R"({"mapBoolBool": {"tr\u0075e": true}})", |
| "map_bool_bool: {key: true value: true}"); |
| |
| // "null" is accepted for all fields types. |
| RunValidJsonTest( |
| "AllFieldAcceptNull", REQUIRED, |
| R"({ |
| "optionalInt32": null, |
| "optionalInt64": null, |
| "optionalUint32": null, |
| "optionalUint64": null, |
| "optionalBool": null, |
| "optionalString": null, |
| "optionalBytes": null, |
| "optionalNestedEnum": null, |
| "optionalNestedMessage": null, |
| "repeatedInt32": null, |
| "repeatedInt64": null, |
| "repeatedUint32": null, |
| "repeatedUint64": null, |
| "repeatedBool": null, |
| "repeatedString": null, |
| "repeatedBytes": null, |
| "repeatedNestedEnum": null, |
| "repeatedNestedMessage": null, |
| "mapInt32Int32": null, |
| "mapBoolBool": null, |
| "mapStringNestedMessage": null |
| })", |
| ""); |
| |
| // Repeated field elements cannot be null. |
| ExpectParseFailureForJson( |
| "RepeatedFieldPrimitiveElementIsNull", RECOMMENDED, |
| R"({"repeatedInt32": [1, null, 2]})"); |
| ExpectParseFailureForJson( |
| "RepeatedFieldMessageElementIsNull", RECOMMENDED, |
| R"({"repeatedNestedMessage": [{"a":1}, null, {"a":2}]})"); |
| // Map field keys cannot be null. |
| ExpectParseFailureForJson( |
| "MapFieldKeyIsNull", RECOMMENDED, |
| R"({"mapInt32Int32": {null: 1}})"); |
| // Map field values cannot be null. |
| ExpectParseFailureForJson( |
| "MapFieldValueIsNull", RECOMMENDED, |
| R"({"mapInt32Int32": {"0": null}})"); |
| |
| // http://www.rfc-editor.org/rfc/rfc7159.txt says strings have to use double |
| // quotes. |
| ExpectParseFailureForJson( |
| "StringFieldSingleQuoteKey", RECOMMENDED, |
| R"({'optionalString': "Hello world!"})"); |
| ExpectParseFailureForJson( |
| "StringFieldSingleQuoteValue", RECOMMENDED, |
| R"({"optionalString": 'Hello world!'})"); |
| ExpectParseFailureForJson( |
| "StringFieldSingleQuoteBoth", RECOMMENDED, |
| R"({'optionalString': 'Hello world!'})"); |
| |
| // Wrapper types. |
| RunValidJsonTest( |
| "OptionalBoolWrapper", REQUIRED, |
| R"({"optionalBoolWrapper": false})", |
| "optional_bool_wrapper: {value: false}"); |
| RunValidJsonTest( |
| "OptionalInt32Wrapper", REQUIRED, |
| R"({"optionalInt32Wrapper": 0})", |
| "optional_int32_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalUint32Wrapper", REQUIRED, |
| R"({"optionalUint32Wrapper": 0})", |
| "optional_uint32_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalInt64Wrapper", REQUIRED, |
| R"({"optionalInt64Wrapper": 0})", |
| "optional_int64_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalUint64Wrapper", REQUIRED, |
| R"({"optionalUint64Wrapper": 0})", |
| "optional_uint64_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalFloatWrapper", REQUIRED, |
| R"({"optionalFloatWrapper": 0})", |
| "optional_float_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalDoubleWrapper", REQUIRED, |
| R"({"optionalDoubleWrapper": 0})", |
| "optional_double_wrapper: {value: 0}"); |
| RunValidJsonTest( |
| "OptionalStringWrapper", REQUIRED, |
| R"({"optionalStringWrapper": ""})", |
| R"(optional_string_wrapper: {value: ""})"); |
| RunValidJsonTest( |
| "OptionalBytesWrapper", REQUIRED, |
| R"({"optionalBytesWrapper": ""})", |
| R"(optional_bytes_wrapper: {value: ""})"); |
| RunValidJsonTest( |
| "OptionalWrapperTypesWithNonDefaultValue", REQUIRED, |
| R"({ |
| "optionalBoolWrapper": true, |
| "optionalInt32Wrapper": 1, |
| "optionalUint32Wrapper": 1, |
| "optionalInt64Wrapper": "1", |
| "optionalUint64Wrapper": "1", |
| "optionalFloatWrapper": 1, |
| "optionalDoubleWrapper": 1, |
| "optionalStringWrapper": "1", |
| "optionalBytesWrapper": "AQI=" |
| })", |
| R"( |
| optional_bool_wrapper: {value: true} |
| optional_int32_wrapper: {value: 1} |
| optional_uint32_wrapper: {value: 1} |
| optional_int64_wrapper: {value: 1} |
| optional_uint64_wrapper: {value: 1} |
| optional_float_wrapper: {value: 1} |
| optional_double_wrapper: {value: 1} |
| optional_string_wrapper: {value: "1"} |
| optional_bytes_wrapper: {value: "\x01\x02"} |
| )"); |
| RunValidJsonTest( |
| "RepeatedBoolWrapper", REQUIRED, |
| R"({"repeatedBoolWrapper": [true, false]})", |
| "repeated_bool_wrapper: {value: true}" |
| "repeated_bool_wrapper: {value: false}"); |
| RunValidJsonTest( |
| "RepeatedInt32Wrapper", REQUIRED, |
| R"({"repeatedInt32Wrapper": [0, 1]})", |
| "repeated_int32_wrapper: {value: 0}" |
| "repeated_int32_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedUint32Wrapper", REQUIRED, |
| R"({"repeatedUint32Wrapper": [0, 1]})", |
| "repeated_uint32_wrapper: {value: 0}" |
| "repeated_uint32_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedInt64Wrapper", REQUIRED, |
| R"({"repeatedInt64Wrapper": [0, 1]})", |
| "repeated_int64_wrapper: {value: 0}" |
| "repeated_int64_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedUint64Wrapper", REQUIRED, |
| R"({"repeatedUint64Wrapper": [0, 1]})", |
| "repeated_uint64_wrapper: {value: 0}" |
| "repeated_uint64_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedFloatWrapper", REQUIRED, |
| R"({"repeatedFloatWrapper": [0, 1]})", |
| "repeated_float_wrapper: {value: 0}" |
| "repeated_float_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedDoubleWrapper", REQUIRED, |
| R"({"repeatedDoubleWrapper": [0, 1]})", |
| "repeated_double_wrapper: {value: 0}" |
| "repeated_double_wrapper: {value: 1}"); |
| RunValidJsonTest( |
| "RepeatedStringWrapper", REQUIRED, |
| R"({"repeatedStringWrapper": ["", "AQI="]})", |
| R"( |
| repeated_string_wrapper: {value: ""} |
| repeated_string_wrapper: {value: "AQI="} |
| )"); |
| RunValidJsonTest( |
| "RepeatedBytesWrapper", REQUIRED, |
| R"({"repeatedBytesWrapper": ["", "AQI="]})", |
| R"( |
| repeated_bytes_wrapper: {value: ""} |
| repeated_bytes_wrapper: {value: "\x01\x02"} |
| )"); |
| RunValidJsonTest( |
| "WrapperTypesWithNullValue", REQUIRED, |
| R"({ |
| "optionalBoolWrapper": null, |
| "optionalInt32Wrapper": null, |
| "optionalUint32Wrapper": null, |
| "optionalInt64Wrapper": null, |
| "optionalUint64Wrapper": null, |
| "optionalFloatWrapper": null, |
| "optionalDoubleWrapper": null, |
| "optionalStringWrapper": null, |
| "optionalBytesWrapper": null, |
| "repeatedBoolWrapper": null, |
| "repeatedInt32Wrapper": null, |
| "repeatedUint32Wrapper": null, |
| "repeatedInt64Wrapper": null, |
| "repeatedUint64Wrapper": null, |
| "repeatedFloatWrapper": null, |
| "repeatedDoubleWrapper": null, |
| "repeatedStringWrapper": null, |
| "repeatedBytesWrapper": null |
| })", |
| ""); |
| |
| // Duration |
| RunValidJsonTest( |
| "DurationMinValue", REQUIRED, |
| R"({"optionalDuration": "-315576000000.999999999s"})", |
| "optional_duration: {seconds: -315576000000 nanos: -999999999}"); |
| RunValidJsonTest( |
| "DurationMaxValue", REQUIRED, |
| R"({"optionalDuration": "315576000000.999999999s"})", |
| "optional_duration: {seconds: 315576000000 nanos: 999999999}"); |
| RunValidJsonTest( |
| "DurationRepeatedValue", REQUIRED, |
| R"({"repeatedDuration": ["1.5s", "-1.5s"]})", |
| "repeated_duration: {seconds: 1 nanos: 500000000}" |
| "repeated_duration: {seconds: -1 nanos: -500000000}"); |
| |
| ExpectParseFailureForJson( |
| "DurationMissingS", REQUIRED, |
| R"({"optionalDuration": "1"})"); |
| ExpectParseFailureForJson( |
| "DurationJsonInputTooSmall", REQUIRED, |
| R"({"optionalDuration": "-315576000001.000000000s"})"); |
| ExpectParseFailureForJson( |
| "DurationJsonInputTooLarge", REQUIRED, |
| R"({"optionalDuration": "315576000001.000000000s"})"); |
| ExpectSerializeFailureForJson( |
| "DurationProtoInputTooSmall", REQUIRED, |
| "optional_duration: {seconds: -315576000001 nanos: 0}"); |
| ExpectSerializeFailureForJson( |
| "DurationProtoInputTooLarge", REQUIRED, |
| "optional_duration: {seconds: 315576000001 nanos: 0}"); |
| |
| RunValidJsonTestWithValidator( |
| "DurationHasZeroFractionalDigit", RECOMMENDED, |
| R"({"optionalDuration": "1.000000000s"})", |
| [](const Json::Value& value) { |
| return value["optionalDuration"].asString() == "1s"; |
| }); |
| RunValidJsonTestWithValidator( |
| "DurationHas3FractionalDigits", RECOMMENDED, |
| R"({"optionalDuration": "1.010000000s"})", |
| [](const Json::Value& value) { |
| return value["optionalDuration"].asString() == "1.010s"; |
| }); |
| RunValidJsonTestWithValidator( |
| "DurationHas6FractionalDigits", RECOMMENDED, |
| R"({"optionalDuration": "1.000010000s"})", |
| [](const Json::Value& value) { |
| return value["optionalDuration"].asString() == "1.000010s"; |
| }); |
| RunValidJsonTestWithValidator( |
| "DurationHas9FractionalDigits", RECOMMENDED, |
| R"({"optionalDuration": "1.000000010s"})", |
| [](const Json::Value& value) { |
| return value["optionalDuration"].asString() == "1.000000010s"; |
| }); |
| |
| // Timestamp |
| RunValidJsonTest( |
| "TimestampMinValue", REQUIRED, |
| R"({"optionalTimestamp": "0001-01-01T00:00:00Z"})", |
| "optional_timestamp: {seconds: -62135596800}"); |
| RunValidJsonTest( |
| "TimestampMaxValue", REQUIRED, |
| R"({"optionalTimestamp": "9999-12-31T23:59:59.999999999Z"})", |
| "optional_timestamp: {seconds: 253402300799 nanos: 999999999}"); |
| RunValidJsonTest( |
| "TimestampRepeatedValue", REQUIRED, |
| R"({ |
| "repeatedTimestamp": [ |
| "0001-01-01T00:00:00Z", |
| "9999-12-31T23:59:59.999999999Z" |
| ] |
| })", |
| "repeated_timestamp: {seconds: -62135596800}" |
| "repeated_timestamp: {seconds: 253402300799 nanos: 999999999}"); |
| RunValidJsonTest( |
| "TimestampWithPositiveOffset", REQUIRED, |
| R"({"optionalTimestamp": "1970-01-01T08:00:00+08:00"})", |
| "optional_timestamp: {seconds: 0}"); |
| RunValidJsonTest( |
| "TimestampWithNegativeOffset", REQUIRED, |
| R"({"optionalTimestamp": "1969-12-31T16:00:00-08:00"})", |
| "optional_timestamp: {seconds: 0}"); |
| |
| ExpectParseFailureForJson( |
| "TimestampJsonInputTooSmall", REQUIRED, |
| R"({"optionalTimestamp": "0000-01-01T00:00:00Z"})"); |
| ExpectParseFailureForJson( |
| "TimestampJsonInputTooLarge", REQUIRED, |
| R"({"optionalTimestamp": "10000-01-01T00:00:00Z"})"); |
| ExpectParseFailureForJson( |
| "TimestampJsonInputMissingZ", REQUIRED, |
| R"({"optionalTimestamp": "0001-01-01T00:00:00"})"); |
| ExpectParseFailureForJson( |
| "TimestampJsonInputMissingT", REQUIRED, |
| R"({"optionalTimestamp": "0001-01-01 00:00:00Z"})"); |
| ExpectParseFailureForJson( |
| "TimestampJsonInputLowercaseZ", REQUIRED, |
| R"({"optionalTimestamp": "0001-01-01T00:00:00z"})"); |
| ExpectParseFailureForJson( |
| "TimestampJsonInputLowercaseT", REQUIRED, |
| R"({"optionalTimestamp": "0001-01-01t00:00:00Z"})"); |
| ExpectSerializeFailureForJson( |
| "TimestampProtoInputTooSmall", REQUIRED, |
| "optional_timestamp: {seconds: -62135596801}"); |
| ExpectSerializeFailureForJson( |
| "TimestampProtoInputTooLarge", REQUIRED, |
| "optional_timestamp: {seconds: 253402300800}"); |
| RunValidJsonTestWithValidator( |
| "TimestampZeroNormalized", RECOMMENDED, |
| R"({"optionalTimestamp": "1969-12-31T16:00:00-08:00"})", |
| [](const Json::Value& value) { |
| return value["optionalTimestamp"].asString() == |
| "1970-01-01T00:00:00Z"; |
| }); |
| RunValidJsonTestWithValidator( |
| "TimestampHasZeroFractionalDigit", RECOMMENDED, |
| R"({"optionalTimestamp": "1970-01-01T00:00:00.000000000Z"})", |
| [](const Json::Value& value) { |
| return value["optionalTimestamp"].asString() == |
| "1970-01-01T00:00:00Z"; |
| }); |
| RunValidJsonTestWithValidator( |
| "TimestampHas3FractionalDigits", RECOMMENDED, |
| R"({"optionalTimestamp": "1970-01-01T00:00:00.010000000Z"})", |
| [](const Json::Value& value) { |
| return value["optionalTimestamp"].asString() == |
| "1970-01-01T00:00:00.010Z"; |
| }); |
| RunValidJsonTestWithValidator( |
| "TimestampHas6FractionalDigits", RECOMMENDED, |
| R"({"optionalTimestamp": "1970-01-01T00:00:00.000010000Z"})", |
| [](const Json::Value& value) { |
| return value["optionalTimestamp"].asString() == |
| "1970-01-01T00:00:00.000010Z"; |
| }); |
| RunValidJsonTestWithValidator( |
| "TimestampHas9FractionalDigits", RECOMMENDED, |
| R"({"optionalTimestamp": "1970-01-01T00:00:00.000000010Z"})", |
| [](const Json::Value& value) { |
| return value["optionalTimestamp"].asString() == |
| "1970-01-01T00:00:00.000000010Z"; |
| }); |
| |
| // FieldMask |
| RunValidJsonTest( |
| "FieldMask", REQUIRED, |
| R"({"optionalFieldMask": "foo,barBaz"})", |
| R"(optional_field_mask: {paths: "foo" paths: "bar_baz"})"); |
| ExpectParseFailureForJson( |
| "FieldMaskInvalidCharacter", RECOMMENDED, |
| R"({"optionalFieldMask": "foo,bar_bar"})"); |
| ExpectSerializeFailureForJson( |
| "FieldMaskPathsDontRoundTrip", RECOMMENDED, |
| R"(optional_field_mask: {paths: "fooBar"})"); |
| ExpectSerializeFailureForJson( |
| "FieldMaskNumbersDontRoundTrip", RECOMMENDED, |
| R"(optional_field_mask: {paths: "foo_3_bar"})"); |
| ExpectSerializeFailureForJson( |
| "FieldMaskTooManyUnderscore", RECOMMENDED, |
| R"(optional_field_mask: {paths: "foo__bar"})"); |
| |
| // Struct |
| RunValidJsonTest( |
| "Struct", REQUIRED, |
| R"({ |
| "optionalStruct": { |
| "nullValue": null, |
| "intValue": 1234, |
| "boolValue": true, |
| "doubleValue": 1234.5678, |
| "stringValue": "Hello world!", |
| "listValue": [1234, "5678"], |
| "objectValue": { |
| "value": 0 |
| } |
| } |
| })", |
| R"( |
| optional_struct: { |
| fields: { |
| key: "nullValue" |
| value: {null_value: NULL_VALUE} |
| } |
| fields: { |
| key: "intValue" |
| value: {number_value: 1234} |
| } |
| fields: { |
| key: "boolValue" |
| value: {bool_value: true} |
| } |
| fields: { |
| key: "doubleValue" |
| value: {number_value: 1234.5678} |
| } |
| fields: { |
| key: "stringValue" |
| value: {string_value: "Hello world!"} |
| } |
| fields: { |
| key: "listValue" |
| value: { |
| list_value: { |
| values: { |
| number_value: 1234 |
| } |
| values: { |
| string_value: "5678" |
| } |
| } |
| } |
| } |
| fields: { |
| key: "objectValue" |
| value: { |
| struct_value: { |
| fields: { |
| key: "value" |
| value: { |
| number_value: 0 |
| } |
| } |
| } |
| } |
| } |
| } |
| )"); |
| // Value |
| RunValidJsonTest( |
| "ValueAcceptInteger", REQUIRED, |
| R"({"optionalValue": 1})", |
| "optional_value: { number_value: 1}"); |
| RunValidJsonTest( |
| "ValueAcceptFloat", REQUIRED, |
| R"({"optionalValue": 1.5})", |
| "optional_value: { number_value: 1.5}"); |
| RunValidJsonTest( |
| "ValueAcceptBool", REQUIRED, |
| R"({"optionalValue": false})", |
| "optional_value: { bool_value: false}"); |
| RunValidJsonTest( |
| "ValueAcceptNull", REQUIRED, |
| R"({"optionalValue": null})", |
| "optional_value: { null_value: NULL_VALUE}"); |
| RunValidJsonTest( |
| "ValueAcceptString", REQUIRED, |
| R"({"optionalValue": "hello"})", |
| R"(optional_value: { string_value: "hello"})"); |
| RunValidJsonTest( |
| "ValueAcceptList", REQUIRED, |
| R"({"optionalValue": [0, "hello"]})", |
| R"( |
| optional_value: { |
| list_value: { |
| values: { |
| number_value: 0 |
| } |
| values: { |
| string_value: "hello" |
| } |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "ValueAcceptObject", REQUIRED, |
| R"({"optionalValue": {"value": 1}})", |
| R"( |
| optional_value: { |
| struct_value: { |
| fields: { |
| key: "value" |
| value: { |
| number_value: 1 |
| } |
| } |
| } |
| } |
| )"); |
| |
| // Any |
| RunValidJsonTest( |
| "Any", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes", |
| "optionalInt32": 12345 |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes] { |
| optional_int32: 12345 |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyNested", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Any", |
| "value": { |
| "@type": "type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes", |
| "optionalInt32": 12345 |
| } |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Any] { |
| [type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes] { |
| optional_int32: 12345 |
| } |
| } |
| } |
| )"); |
| // The special "@type" tag is not required to appear first. |
| RunValidJsonTest( |
| "AnyUnorderedTypeTag", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "optionalInt32": 12345, |
| "@type": "type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes" |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/protobuf_test_messages.proto3.TestAllTypes] { |
| optional_int32: 12345 |
| } |
| } |
| )"); |
| // Well-known types in Any. |
| RunValidJsonTest( |
| "AnyWithInt32ValueWrapper", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Int32Value", |
| "value": 12345 |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Int32Value] { |
| value: 12345 |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithDuration", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Duration", |
| "value": "1.5s" |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Duration] { |
| seconds: 1 |
| nanos: 500000000 |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithTimestamp", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Timestamp", |
| "value": "1970-01-01T00:00:00Z" |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Timestamp] { |
| seconds: 0 |
| nanos: 0 |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithFieldMask", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.FieldMask", |
| "value": "foo,barBaz" |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.FieldMask] { |
| paths: ["foo", "bar_baz"] |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithStruct", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Struct", |
| "value": { |
| "foo": 1 |
| } |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Struct] { |
| fields: { |
| key: "foo" |
| value: { |
| number_value: 1 |
| } |
| } |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithValueForJsonObject", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Value", |
| "value": { |
| "foo": 1 |
| } |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Value] { |
| struct_value: { |
| fields: { |
| key: "foo" |
| value: { |
| number_value: 1 |
| } |
| } |
| } |
| } |
| } |
| )"); |
| RunValidJsonTest( |
| "AnyWithValueForInteger", REQUIRED, |
| R"({ |
| "optionalAny": { |
| "@type": "type.googleapis.com/google.protobuf.Value", |
| "value": 1 |
| } |
| })", |
| R"( |
| optional_any: { |
| [type.googleapis.com/google.protobuf.Value] { |
| number_value: 1 |
| } |
| } |
| )"); |
| |
| bool ok = true; |
| if (!CheckSetEmpty(expected_to_fail_, "nonexistent_tests.txt", |
| "These tests were listed in the failure list, but they " |
| "don't exist. Remove them from the failure list by " |
| "running:\n" |
| " ./update_failure_list.py " + failure_list_filename_ + |
| " --remove nonexistent_tests.txt")) { |
| ok = false; |
| } |
| if (!CheckSetEmpty(unexpected_failing_tests_, "failing_tests.txt", |
| "These tests failed. If they can't be fixed right now, " |
| "you can add them to the failure list so the overall " |
| "suite can succeed. Add them to the failure list by " |
| "running:\n" |
| " ./update_failure_list.py " + failure_list_filename_ + |
| " --add failing_tests.txt")) { |
| ok = false; |
| } |
| if (!CheckSetEmpty(unexpected_succeeding_tests_, "succeeding_tests.txt", |
| "These tests succeeded, even though they were listed in " |
| "the failure list. Remove them from the failure list " |
| "by running:\n" |
| " ./update_failure_list.py " + failure_list_filename_ + |
| " --remove succeeding_tests.txt")) { |
| ok = false; |
| } |
| |
| if (verbose_) { |
| CheckSetEmpty(skipped_, "", |
| "These tests were skipped (probably because support for some " |
| "features is not implemented)"); |
| } |
| |
| StringAppendF(&output_, |
| "CONFORMANCE SUITE %s: %d successes, %d skipped, " |
| "%d expected failures, %d unexpected failures.\n", |
| ok ? "PASSED" : "FAILED", successes_, skipped_.size(), |
| expected_failures_, unexpected_failing_tests_.size()); |
| StringAppendF(&output_, "\n"); |
| |
| output->assign(output_); |
| |
| return ok; |
| } |
| |
| } // namespace protobuf |
| } // namespace google |