blob: 7a8aea8f72a4bcd85811b7050ddae660765157ce [file] [log] [blame]
// Copyright (c) 2006-2008 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
// This command-line program converts an effective-TLD data file in UTF-8 from
// the format provided by Mozilla to the format expected by Chrome. This
// program generates an intermediate file which is then used by gperf to
// generate a perfect hash map. The benefit of this approach is that no time is
// spent on program initialization to generate the map of this data.
//
// Running this program finds "effective_tld_names.cc" in the expected location
// in the source checkout and generates "effective_tld_names.gperf" next to it.
//
// Any errors or warnings from this program are recorded in tld_cleanup.log.
//
// In particular, it
// * Strips blank lines and comments, as well as notes for individual rules.
// * Strips a single leading and/or trailing dot from each rule, if present.
// * Logs a warning if a rule contains '!' or '*.' other than at the beginning
// of the rule. (This also catches multiple ! or *. at the start of a rule.)
// * Logs a warning if GURL reports a rule as invalid, but keeps the rule.
// * Canonicalizes each rule's domain by converting it to a GURL and back.
// * Adds explicit rules for true TLDs found in any rule.
#include <map>
#include <set>
#include <string>
#include "base/at_exit.h"
#include "base/command_line.h"
#include "base/file_util.h"
#include "base/i18n/icu_util.h"
#include "base/logging.h"
#include "base/file_path.h"
#include "base/file_util.h"
#include "base/path_service.h"
#include "base/process_util.h"
#include "base/string_util.h"
#include "googleurl/src/gurl.h"
#include "googleurl/src/url_parse.h"
namespace {
struct Rule {
bool exception;
bool wildcard;
};
typedef std::map<std::string, Rule> RuleMap;
typedef std::set<std::string> RuleSet;
}
// Writes the list of domain rules contained in the 'rules' set to the
// 'outfile', with each rule terminated by a LF. The file must already have
// been created with write access.
bool WriteRules(const RuleMap& rules, FilePath outfile) {
std::string data;
data.append(
"%{\n"
"// Copyright (c) 2009 The Chromium Authors. All rights reserved.\n"
"// Use of this source code is governed by a BSD-style license that\n"
"// can be found in the LICENSE file.\n\n"
"// This file is generated by net/tools/tld_cleanup/.\n"
"// DO NOT MANUALLY EDIT!\n"
"%}\n"
"struct DomainRule {\n"
" const char *name;\n"
" int type; // 1: exception, 2: wildcard\n"
"};\n"
"%%\n"
);
for (RuleMap::const_iterator i = rules.begin(); i != rules.end(); ++i) {
data.append(i->first);
data.append(", ");
if (i->second.exception) {
data.append("1");
} else if (i->second.wildcard) {
data.append("2");
} else {
data.append("0");
}
data.append("\n");
}
data.append("%%\n");
int written = file_util::WriteFile(outfile, data.data(), data.size());
return written == static_cast<int>(data.size());
}
// These result codes should be in increasing order of severity.
typedef enum {
kSuccess,
kWarning,
kError,
} NormalizeResult;
// Adjusts the rule to a standard form: removes single extraneous dots and
// canonicalizes it using GURL. Returns kSuccess if the rule is interpreted as
// valid; logs a warning and returns kWarning if it is probably invalid; and
// logs an error and returns kError if the rule is (almost) certainly invalid.
NormalizeResult NormalizeRule(std::string* domain, Rule* rule) {
NormalizeResult result = kSuccess;
// Strip single leading and trailing dots.
if (domain->at(0) == '.')
domain->erase(0, 1);
if (domain->empty()) {
LOG(WARNING) << "Ignoring empty rule";
return kWarning;
}
if (domain->at(domain->size() - 1) == '.')
domain->erase(domain->size() - 1, 1);
if (domain->empty()) {
LOG(WARNING) << "Ignoring empty rule";
return kWarning;
}
// Allow single leading '*.' or '!', saved here so it's not canonicalized.
size_t start_offset = 0;
if (domain->at(0) == '!') {
domain->erase(0, 1);
rule->exception = true;
} else if (domain->find("*.") == 0) {
domain->erase(0, 2);
rule->wildcard = true;
}
if (domain->empty()) {
LOG(WARNING) << "Ignoring empty rule";
return kWarning;
}
// Warn about additional '*.' or '!'.
if (domain->find("*.", start_offset) != std::string::npos ||
domain->find('!', start_offset) != std::string::npos) {
LOG(WARNING) << "Keeping probably invalid rule: " << *domain;
result = kWarning;
}
// Make a GURL and normalize it, then get the host back out.
std::string url = "http://";
url.append(*domain);
GURL gurl(url);
const std::string& spec = gurl.possibly_invalid_spec();
url_parse::Component host = gurl.parsed_for_possibly_invalid_spec().host;
if (host.len < 0) {
LOG(ERROR) << "Ignoring rule that couldn't be normalized: " << *domain;
return kError;
}
if (!gurl.is_valid()) {
LOG(WARNING) << "Keeping rule that GURL says is invalid: " << *domain;
result = kWarning;
}
domain->assign(spec.substr(host.begin, host.len));
return result;
}
// Loads the file described by 'in_filename', converts it to the desired format
// (see the file comments above), and saves it into 'out_filename'. Returns
// the most severe of the result codes encountered when normalizing the rules.
NormalizeResult NormalizeFile(const FilePath& in_filename,
const FilePath& out_filename) {
std::string data;
if (!file_util::ReadFileToString(in_filename, &data)) {
LOG(ERROR) << "Unable to read file";
// We return success since we've already reported the error.
return kSuccess;
}
// We do a lot of string assignment during parsing, but simplicity is more
// important than performance here.
std::string domain;
NormalizeResult result = kSuccess;
size_t line_start = 0;
size_t line_end = 0;
RuleMap rules;
RuleSet extra_rules;
while (line_start < data.size()) {
// Skip comments.
if (line_start + 1 < data.size() &&
data[line_start] == '/' &&
data[line_start + 1] == '/') {
line_end = data.find_first_of("\r\n", line_start);
if (line_end == std::string::npos)
line_end = data.size();
} else {
// Truncate at first whitespace.
line_end = data.find_first_of("\r\n \t", line_start);
if (line_end == std::string::npos)
line_end = data.size();
domain.assign(data.data(), line_start, line_end - line_start);
Rule rule;
rule.wildcard = false;
rule.exception = false;
NormalizeResult new_result = NormalizeRule(&domain, &rule);
if (new_result != kError) {
// Check the existing rules to make sure we don't have an exception and
// wildcard for the same rule. If we did, we'd have to update our
// parsing code to handle this case.
CHECK(rules.find(domain) == rules.end());
rules[domain] = rule;
// Add true TLD for multi-level rules. We don't add them right now, in
// case there's an exception or wild card that either exists or might be
// added in a later iteration. In those cases, there's no need to add
// it and it would just slow down parsing the data.
size_t tld_start = domain.find_last_of('.');
if (tld_start != std::string::npos && tld_start + 1 < domain.size())
extra_rules.insert(domain.substr(tld_start + 1));
}
result = std::max(result, new_result);
}
// Find beginning of next non-empty line.
line_start = data.find_first_of("\r\n", line_end);
if (line_start == std::string::npos)
line_start = data.size();
line_start = data.find_first_not_of("\r\n", line_start);
if (line_start == std::string::npos)
line_start = data.size();
}
for (RuleSet::const_iterator iter = extra_rules.begin();
iter != extra_rules.end();
++iter) {
if (rules.find(*iter) == rules.end()) {
Rule rule;
rule.exception = false;
rule.wildcard = false;
rules[*iter] = rule;
}
}
if (!WriteRules(rules, out_filename)) {
LOG(ERROR) << "Error(s) writing output file";
result = kError;
}
return result;
}
int main(int argc, const char* argv[]) {
base::EnableTerminationOnHeapCorruption();
if (argc != 1) {
fprintf(stderr, "Normalizes and verifies UTF-8 TLD data files\n");
fprintf(stderr, "Usage: %s\n", argv[0]);
return 1;
}
// Manages the destruction of singletons.
base::AtExitManager exit_manager;
// Only use OutputDebugString in debug mode.
#ifdef NDEBUG
logging::LoggingDestination destination = logging::LOG_ONLY_TO_FILE;
#else
logging::LoggingDestination destination =
logging::LOG_TO_BOTH_FILE_AND_SYSTEM_DEBUG_LOG;
#endif
CommandLine::Init(argc, argv);
FilePath log_filename;
PathService::Get(base::DIR_EXE, &log_filename);
log_filename = log_filename.AppendASCII("tld_cleanup.log");
logging::InitLogging(
log_filename.value().c_str(),
destination,
logging::LOCK_LOG_FILE,
logging::DELETE_OLD_LOG_FILE,
logging::DISABLE_DCHECK_FOR_NON_OFFICIAL_RELEASE_BUILDS);
icu_util::Initialize();
FilePath input_file;
PathService::Get(base::DIR_SOURCE_ROOT, &input_file);
input_file = input_file.Append(FILE_PATH_LITERAL("net"))
.Append(FILE_PATH_LITERAL("base"))
.Append(FILE_PATH_LITERAL("effective_tld_names.dat"));
FilePath output_file;
PathService::Get(base::DIR_SOURCE_ROOT, &output_file);
output_file = output_file.Append(FILE_PATH_LITERAL("net"))
.Append(FILE_PATH_LITERAL("base"))
.Append(FILE_PATH_LITERAL(
"effective_tld_names.gperf"));
NormalizeResult result = NormalizeFile(input_file, output_file);
if (result != kSuccess) {
fprintf(stderr,
"Errors or warnings processing file. See log in tld_cleanup.log.");
}
if (result == kError)
return 1;
return 0;
}