| // Copyright (c) 2010 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "chrome/browser/safe_browsing/protocol_manager.h" |
| |
| #ifndef NDEBUG |
| #include "base/base64.h" |
| #endif |
| #include "base/environment.h" |
| #include "base/logging.h" |
| #include "base/metrics/histogram.h" |
| #include "base/rand_util.h" |
| #include "base/stl_util-inl.h" |
| #include "base/string_util.h" |
| #include "base/task.h" |
| #include "base/timer.h" |
| #include "chrome/browser/browser_thread.h" |
| #include "chrome/browser/safe_browsing/protocol_parser.h" |
| #include "chrome/browser/safe_browsing/safe_browsing_service.h" |
| #include "chrome/common/chrome_version_info.h" |
| #include "chrome/common/env_vars.h" |
| #include "chrome/common/net/url_request_context_getter.h" |
| #include "net/base/escape.h" |
| #include "net/base/load_flags.h" |
| #include "net/url_request/url_request_status.h" |
| |
| using base::Time; |
| using base::TimeDelta; |
| |
| // Maximum time, in seconds, from start up before we must issue an update query. |
| static const int kSbTimerStartIntervalSec = 5 * 60; |
| |
| // The maximum time, in seconds, to wait for a response to an update request. |
| static const int kSbMaxUpdateWaitSec = 10; |
| |
| // Maximum back off multiplier. |
| static const int kSbMaxBackOff = 8; |
| |
| // The default SBProtocolManagerFactory. |
| class SBProtocolManagerFactoryImpl : public SBProtocolManagerFactory { |
| public: |
| SBProtocolManagerFactoryImpl() { } |
| virtual ~SBProtocolManagerFactoryImpl() { } |
| virtual SafeBrowsingProtocolManager* CreateProtocolManager( |
| SafeBrowsingService* sb_service, |
| const std::string& client_name, |
| const std::string& client_key, |
| const std::string& wrapped_key, |
| URLRequestContextGetter* request_context_getter, |
| const std::string& info_url_prefix, |
| const std::string& mackey_url_prefix, |
| bool disable_auto_update) { |
| return new SafeBrowsingProtocolManager( |
| sb_service, client_name, client_key, wrapped_key, |
| request_context_getter, info_url_prefix, mackey_url_prefix, |
| disable_auto_update); |
| } |
| private: |
| DISALLOW_COPY_AND_ASSIGN(SBProtocolManagerFactoryImpl); |
| }; |
| |
| // SafeBrowsingProtocolManager implementation ---------------------------------- |
| |
| // static |
| SBProtocolManagerFactory* SafeBrowsingProtocolManager::factory_ = NULL; |
| |
| // static |
| SafeBrowsingProtocolManager* SafeBrowsingProtocolManager::Create( |
| SafeBrowsingService* sb_service, |
| const std::string& client_name, |
| const std::string& client_key, |
| const std::string& wrapped_key, |
| URLRequestContextGetter* request_context_getter, |
| const std::string& info_url_prefix, |
| const std::string& mackey_url_prefix, |
| bool disable_auto_update) { |
| DCHECK(BrowserThread::CurrentlyOn(BrowserThread::IO)); |
| if (!factory_) |
| factory_ = new SBProtocolManagerFactoryImpl(); |
| return factory_->CreateProtocolManager(sb_service, client_name, client_key, |
| wrapped_key, request_context_getter, |
| info_url_prefix, mackey_url_prefix, |
| disable_auto_update); |
| } |
| |
| SafeBrowsingProtocolManager::SafeBrowsingProtocolManager( |
| SafeBrowsingService* sb_service, |
| const std::string& client_name, |
| const std::string& client_key, |
| const std::string& wrapped_key, |
| URLRequestContextGetter* request_context_getter, |
| const std::string& http_url_prefix, |
| const std::string& https_url_prefix, |
| bool disable_auto_update) |
| : sb_service_(sb_service), |
| request_type_(NO_REQUEST), |
| update_error_count_(0), |
| gethash_error_count_(0), |
| update_back_off_mult_(1), |
| gethash_back_off_mult_(1), |
| next_update_sec_(-1), |
| update_state_(FIRST_REQUEST), |
| initial_request_(true), |
| chunk_pending_to_write_(false), |
| client_key_(client_key), |
| wrapped_key_(wrapped_key), |
| update_size_(0), |
| client_name_(client_name), |
| request_context_getter_(request_context_getter), |
| http_url_prefix_(http_url_prefix), |
| https_url_prefix_(https_url_prefix), |
| disable_auto_update_(disable_auto_update) { |
| DCHECK(!http_url_prefix_.empty() && !https_url_prefix_.empty()); |
| |
| // Set the backoff multiplier fuzz to a random value between 0 and 1. |
| back_off_fuzz_ = static_cast<float>(base::RandDouble()); |
| // The first update must happen between 1-5 minutes of start up. |
| next_update_sec_ = base::RandInt(60, kSbTimerStartIntervalSec); |
| |
| chrome::VersionInfo version_info; |
| if (!version_info.is_valid() || version_info.Version().empty()) |
| version_ = "0.1"; |
| else |
| version_ = version_info.Version(); |
| } |
| |
| // static |
| void SafeBrowsingProtocolManager::RecordGetHashResult(ResultType result_type) { |
| UMA_HISTOGRAM_ENUMERATION("SB2.GetHashResult", result_type, |
| GET_HASH_RESULT_MAX); |
| } |
| |
| SafeBrowsingProtocolManager::~SafeBrowsingProtocolManager() { |
| // Delete in-progress SafeBrowsing requests. |
| STLDeleteContainerPairFirstPointers(hash_requests_.begin(), |
| hash_requests_.end()); |
| hash_requests_.clear(); |
| |
| // Delete in-progress safebrowsing reports (hits and details). |
| STLDeleteContainerPointers(safebrowsing_reports_.begin(), |
| safebrowsing_reports_.end()); |
| safebrowsing_reports_.clear(); |
| } |
| |
| // Public API used by the SafeBrowsingService ---------------------------------- |
| |
| // We can only have one update or chunk request outstanding, but there may be |
| // multiple GetHash requests pending since we don't want to serialize them and |
| // slow down the user. |
| void SafeBrowsingProtocolManager::GetFullHash( |
| SafeBrowsingService::SafeBrowsingCheck* check, |
| const std::vector<SBPrefix>& prefixes) { |
| // If we are in GetHash backoff, we need to check if we're past the next |
| // allowed time. If we are, we can proceed with the request. If not, we are |
| // required to return empty results (i.e. treat the page as safe). |
| if (gethash_error_count_ && Time::Now() <= next_gethash_time_) { |
| std::vector<SBFullHashResult> full_hashes; |
| sb_service_->HandleGetHashResults(check, full_hashes, false); |
| return; |
| } |
| bool use_mac = !client_key_.empty(); |
| GURL gethash_url = GetHashUrl(use_mac); |
| URLFetcher* fetcher = new URLFetcher(gethash_url, URLFetcher::POST, this); |
| hash_requests_[fetcher] = check; |
| |
| std::string get_hash; |
| SafeBrowsingProtocolParser parser; |
| parser.FormatGetHash(prefixes, &get_hash); |
| |
| fetcher->set_load_flags(net::LOAD_DISABLE_CACHE); |
| fetcher->set_request_context(request_context_getter_); |
| fetcher->set_upload_data("text/plain", get_hash); |
| fetcher->Start(); |
| } |
| |
| void SafeBrowsingProtocolManager::GetNextUpdate() { |
| if (initial_request_) { |
| if (client_key_.empty() || wrapped_key_.empty()) { |
| IssueKeyRequest(); |
| return; |
| } else { |
| initial_request_ = false; |
| } |
| } |
| |
| if (!request_.get()) |
| IssueUpdateRequest(); |
| } |
| |
| // URLFetcher::Delegate implementation ----------------------------------------- |
| |
| // All SafeBrowsing request responses are handled here. |
| // TODO(paulg): Clarify with the SafeBrowsing team whether a failed parse of a |
| // chunk should retry the download and parse of that chunk (and |
| // what back off / how many times to try), and if that effects the |
| // update back off. For now, a failed parse of the chunk means we |
| // drop it. This isn't so bad because the next UPDATE_REQUEST we |
| // do will report all the chunks we have. If that chunk is still |
| // required, the SafeBrowsing servers will tell us to get it again. |
| void SafeBrowsingProtocolManager::OnURLFetchComplete( |
| const URLFetcher* source, |
| const GURL& url, |
| const URLRequestStatus& status, |
| int response_code, |
| const ResponseCookies& cookies, |
| const std::string& data) { |
| scoped_ptr<const URLFetcher> fetcher; |
| bool parsed_ok = true; |
| bool must_back_off = false; // Reduce SafeBrowsing service query frequency. |
| |
| // See if this is a safebrowsing report fetcher. We don't take any action for |
| // the response to those. |
| std::set<const URLFetcher*>::iterator sit = safebrowsing_reports_.find( |
| source); |
| if (sit != safebrowsing_reports_.end()) { |
| const URLFetcher* report = *sit; |
| safebrowsing_reports_.erase(sit); |
| delete report; |
| return; |
| } |
| |
| HashRequests::iterator it = hash_requests_.find(source); |
| if (it != hash_requests_.end()) { |
| // GetHash response. |
| fetcher.reset(it->first); |
| SafeBrowsingService::SafeBrowsingCheck* check = it->second; |
| std::vector<SBFullHashResult> full_hashes; |
| bool can_cache = false; |
| if (response_code == 200 || response_code == 204) { |
| // For tracking our GetHash false positive (204) rate, compared to real |
| // (200) responses. |
| if (response_code == 200) |
| RecordGetHashResult(GET_HASH_STATUS_200); |
| else |
| RecordGetHashResult(GET_HASH_STATUS_204); |
| can_cache = true; |
| gethash_error_count_ = 0; |
| gethash_back_off_mult_ = 1; |
| bool re_key = false; |
| SafeBrowsingProtocolParser parser; |
| parsed_ok = parser.ParseGetHash(data.data(), |
| static_cast<int>(data.length()), |
| client_key_, |
| &re_key, |
| &full_hashes); |
| if (!parsed_ok) { |
| // If we fail to parse it, we must still inform the SafeBrowsingService |
| // so that it doesn't hold up the user's request indefinitely. Not sure |
| // what to do at that point though! |
| full_hashes.clear(); |
| } else { |
| if (re_key) |
| HandleReKey(); |
| } |
| } else { |
| HandleGetHashError(Time::Now()); |
| if (status.status() == URLRequestStatus::FAILED) { |
| VLOG(1) << "SafeBrowsing GetHash request for: " << source->url() |
| << " failed with os error: " << status.os_error(); |
| } else { |
| VLOG(1) << "SafeBrowsing GetHash request for: " << source->url() |
| << " failed with error: " << response_code; |
| } |
| } |
| |
| // Call back the SafeBrowsingService with full_hashes, even if there was a |
| // parse error or an error response code (in which case full_hashes will be |
| // empty). We can't block the user regardless of the error status. |
| sb_service_->HandleGetHashResults(check, full_hashes, can_cache); |
| |
| hash_requests_.erase(it); |
| } else { |
| // Update, chunk or key response. |
| fetcher.reset(request_.release()); |
| |
| if (request_type_ == UPDATE_REQUEST) { |
| if (!fetcher.get()) { |
| // We've timed out waiting for an update response, so we've cancelled |
| // the update request and scheduled a new one. Ignore this response. |
| return; |
| } |
| |
| // Cancel the update response timeout now that we have the response. |
| update_timer_.Stop(); |
| } |
| |
| if (response_code == 200) { |
| // We have data from the SafeBrowsing service. |
| parsed_ok = HandleServiceResponse(source->url(), |
| data.data(), |
| static_cast<int>(data.length())); |
| if (!parsed_ok) { |
| VLOG(1) << "SafeBrowsing request for: " << source->url() |
| << " failed parse."; |
| must_back_off = true; |
| chunk_request_urls_.clear(); |
| UpdateFinished(false); |
| } |
| |
| switch (request_type_) { |
| case CHUNK_REQUEST: |
| if (parsed_ok) |
| chunk_request_urls_.pop_front(); |
| break; |
| case GETKEY_REQUEST: |
| if (initial_request_) { |
| // This is the first request we've made this session. Now that we |
| // have the keys, do the regular update request. |
| initial_request_ = false; |
| GetNextUpdate(); |
| return; |
| } |
| break; |
| case UPDATE_REQUEST: |
| if (chunk_request_urls_.empty() && parsed_ok) { |
| // We are up to date since the servers gave us nothing new, so we |
| // are done with this update cycle. |
| UpdateFinished(true); |
| } |
| break; |
| default: |
| NOTREACHED(); |
| break; |
| } |
| } else { |
| // The SafeBrowsing service error, or very bad response code: back off. |
| must_back_off = true; |
| if (request_type_ == CHUNK_REQUEST) |
| chunk_request_urls_.clear(); |
| UpdateFinished(false); |
| if (status.status() == URLRequestStatus::FAILED) { |
| VLOG(1) << "SafeBrowsing request for: " << source->url() |
| << " failed with os error: " << status.os_error(); |
| } else { |
| VLOG(1) << "SafeBrowsing request for: " << source->url() |
| << " failed with error: " << response_code; |
| } |
| } |
| } |
| |
| // Schedule a new update request if we've finished retrieving all the chunks |
| // from the previous update. We treat the update request and the chunk URLs it |
| // contains as an atomic unit as far as back off is concerned. |
| if (chunk_request_urls_.empty() && |
| (request_type_ == CHUNK_REQUEST || request_type_ == UPDATE_REQUEST)) |
| ScheduleNextUpdate(must_back_off); |
| |
| // Get the next chunk if available. |
| IssueChunkRequest(); |
| } |
| |
| bool SafeBrowsingProtocolManager::HandleServiceResponse(const GURL& url, |
| const char* data, |
| int length) { |
| SafeBrowsingProtocolParser parser; |
| |
| switch (request_type_) { |
| case UPDATE_REQUEST: { |
| int next_update_sec = -1; |
| bool re_key = false; |
| bool reset = false; |
| scoped_ptr<std::vector<SBChunkDelete> > chunk_deletes( |
| new std::vector<SBChunkDelete>); |
| std::vector<ChunkUrl> chunk_urls; |
| if (!parser.ParseUpdate(data, length, client_key_, |
| &next_update_sec, &re_key, |
| &reset, chunk_deletes.get(), &chunk_urls)) { |
| return false; |
| } |
| |
| last_update_ = Time::Now(); |
| |
| if (update_state_ == FIRST_REQUEST) |
| update_state_ = SECOND_REQUEST; |
| else if (update_state_ == SECOND_REQUEST) |
| update_state_ = NORMAL_REQUEST; |
| |
| // New time for the next update. |
| if (next_update_sec > 0) { |
| next_update_sec_ = next_update_sec; |
| } else if (update_state_ == SECOND_REQUEST) { |
| next_update_sec_ = base::RandInt(15 * 60, 45 * 60); |
| } |
| |
| // We need to request a new set of keys for MAC. |
| if (re_key) |
| HandleReKey(); |
| |
| // New chunks to download. |
| if (!chunk_urls.empty()) { |
| UMA_HISTOGRAM_COUNTS("SB2.UpdateUrls", chunk_urls.size()); |
| for (size_t i = 0; i < chunk_urls.size(); ++i) |
| chunk_request_urls_.push_back(chunk_urls[i]); |
| } |
| |
| // Handle the case were the SafeBrowsing service tells us to dump our |
| // database. |
| if (reset) { |
| sb_service_->ResetDatabase(); |
| return true; |
| } |
| |
| // Chunks to delete from our storage. Pass ownership of |
| // |chunk_deletes|. |
| if (!chunk_deletes->empty()) |
| sb_service_->HandleChunkDelete(chunk_deletes.release()); |
| |
| break; |
| } |
| case CHUNK_REQUEST: { |
| UMA_HISTOGRAM_TIMES("SB2.ChunkRequest", |
| base::Time::Now() - chunk_request_start_); |
| |
| const ChunkUrl chunk_url = chunk_request_urls_.front(); |
| bool re_key = false; |
| scoped_ptr<SBChunkList> chunks(new SBChunkList); |
| UMA_HISTOGRAM_COUNTS("SB2.ChunkSize", length); |
| update_size_ += length; |
| if (!parser.ParseChunk(data, length, |
| client_key_, chunk_url.mac, |
| &re_key, chunks.get())) { |
| #ifndef NDEBUG |
| std::string data_str; |
| data_str.assign(data, length); |
| std::string encoded_chunk; |
| base::Base64Encode(data, &encoded_chunk); |
| VLOG(1) << "ParseChunk error for chunk: " << chunk_url.url |
| << ", client_key: " << client_key_ |
| << ", wrapped_key: " << wrapped_key_ |
| << ", mac: " << chunk_url.mac |
| << ", Base64Encode(data): " << encoded_chunk |
| << ", length: " << length; |
| #endif |
| return false; |
| } |
| |
| if (re_key) |
| HandleReKey(); |
| |
| // Chunks to add to storage. Pass ownership of |chunks|. |
| if (!chunks->empty()) { |
| chunk_pending_to_write_ = true; |
| sb_service_->HandleChunk(chunk_url.list_name, chunks.release()); |
| } |
| |
| break; |
| } |
| case GETKEY_REQUEST: { |
| std::string client_key, wrapped_key; |
| if (!parser.ParseNewKey(data, length, &client_key, &wrapped_key)) |
| return false; |
| |
| client_key_ = client_key; |
| wrapped_key_ = wrapped_key; |
| BrowserThread::PostTask( |
| BrowserThread::UI, FROM_HERE, |
| NewRunnableMethod( |
| sb_service_, &SafeBrowsingService::OnNewMacKeys, client_key_, |
| wrapped_key_)); |
| break; |
| } |
| |
| default: |
| return false; |
| } |
| |
| return true; |
| } |
| |
| void SafeBrowsingProtocolManager::Initialize() { |
| // Don't want to hit the safe browsing servers on build/chrome bots. |
| scoped_ptr<base::Environment> env(base::Environment::Create()); |
| if (env->HasVar(env_vars::kHeadless)) |
| return; |
| |
| ScheduleNextUpdate(false /* no back off */); |
| } |
| |
| void SafeBrowsingProtocolManager::ScheduleNextUpdate(bool back_off) { |
| DCHECK_GT(next_update_sec_, 0); |
| |
| if (disable_auto_update_) { |
| // Unschedule any current timer. |
| update_timer_.Stop(); |
| return; |
| } |
| // Reschedule with the new update. |
| const int next_update = GetNextUpdateTime(back_off); |
| ForceScheduleNextUpdate(next_update); |
| } |
| |
| void SafeBrowsingProtocolManager::ForceScheduleNextUpdate( |
| const int next_update_msec) { |
| DCHECK_GE(next_update_msec, 0); |
| // Unschedule any current timer. |
| update_timer_.Stop(); |
| update_timer_.Start(TimeDelta::FromMilliseconds(next_update_msec), this, |
| &SafeBrowsingProtocolManager::GetNextUpdate); |
| } |
| |
| // According to section 5 of the SafeBrowsing protocol specification, we must |
| // back off after a certain number of errors. We only change 'next_update_sec_' |
| // when we receive a response from the SafeBrowsing service. |
| int SafeBrowsingProtocolManager::GetNextUpdateTime(bool back_off) { |
| int next = next_update_sec_; |
| if (back_off) { |
| next = GetNextBackOffTime(&update_error_count_, &update_back_off_mult_); |
| } else { |
| // Successful response means error reset. |
| update_error_count_ = 0; |
| update_back_off_mult_ = 1; |
| } |
| return next * 1000; // milliseconds |
| } |
| |
| int SafeBrowsingProtocolManager::GetNextBackOffTime(int* error_count, |
| int* multiplier) { |
| DCHECK(multiplier && error_count); |
| (*error_count)++; |
| if (*error_count > 1 && *error_count < 6) { |
| int next = static_cast<int>(*multiplier * (1 + back_off_fuzz_) * 30 * 60); |
| *multiplier *= 2; |
| if (*multiplier > kSbMaxBackOff) |
| *multiplier = kSbMaxBackOff; |
| return next; |
| } |
| |
| if (*error_count >= 6) |
| return 60 * 60 * 8; // 8 hours |
| |
| return 60; // 1 minute |
| } |
| |
| // This request requires getting a list of all the chunks for each list from the |
| // database asynchronously. The request will be issued when we're called back in |
| // OnGetChunksComplete. |
| // TODO(paulg): We should get this at start up and maintain a ChunkRange cache |
| // to avoid hitting the database with each update request. On the |
| // otherhand, this request will only occur ~20-30 minutes so there |
| // isn't that much overhead. Measure! |
| void SafeBrowsingProtocolManager::IssueUpdateRequest() { |
| request_type_ = UPDATE_REQUEST; |
| sb_service_->UpdateStarted(); |
| } |
| |
| void SafeBrowsingProtocolManager::IssueChunkRequest() { |
| // We are only allowed to have one request outstanding at any time. Also, |
| // don't get the next url until the previous one has been written to disk so |
| // that we don't use too much memory. |
| if (request_.get() || chunk_request_urls_.empty() || chunk_pending_to_write_) |
| return; |
| |
| ChunkUrl next_chunk = chunk_request_urls_.front(); |
| DCHECK(!next_chunk.url.empty()); |
| GURL chunk_url = NextChunkUrl(next_chunk.url); |
| request_type_ = CHUNK_REQUEST; |
| request_.reset(new URLFetcher(chunk_url, URLFetcher::GET, this)); |
| request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
| request_->set_request_context(request_context_getter_); |
| chunk_request_start_ = base::Time::Now(); |
| request_->Start(); |
| } |
| |
| void SafeBrowsingProtocolManager::IssueKeyRequest() { |
| GURL key_url = MacKeyUrl(); |
| request_type_ = GETKEY_REQUEST; |
| request_.reset(new URLFetcher(key_url, URLFetcher::GET, this)); |
| request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
| request_->set_request_context(request_context_getter_); |
| request_->Start(); |
| } |
| |
| void SafeBrowsingProtocolManager::OnGetChunksComplete( |
| const std::vector<SBListChunkRanges>& lists, bool database_error) { |
| DCHECK_EQ(request_type_, UPDATE_REQUEST); |
| if (database_error) { |
| UpdateFinished(false); |
| ScheduleNextUpdate(false); |
| return; |
| } |
| |
| const bool use_mac = !client_key_.empty(); |
| |
| // Format our stored chunks: |
| std::string list_data; |
| bool found_malware = false; |
| bool found_phishing = false; |
| for (size_t i = 0; i < lists.size(); ++i) { |
| list_data.append(FormatList(lists[i], use_mac)); |
| if (lists[i].name == safe_browsing_util::kPhishingList) |
| found_phishing = true; |
| |
| if (lists[i].name == safe_browsing_util::kMalwareList) |
| found_malware = true; |
| } |
| |
| // If we have an empty database, let the server know we want data for these |
| // lists. |
| if (!found_phishing) |
| list_data.append(FormatList( |
| SBListChunkRanges(safe_browsing_util::kPhishingList), use_mac)); |
| |
| if (!found_malware) |
| list_data.append(FormatList( |
| SBListChunkRanges(safe_browsing_util::kMalwareList), use_mac)); |
| |
| GURL update_url = UpdateUrl(use_mac); |
| request_.reset(new URLFetcher(update_url, URLFetcher::POST, this)); |
| request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
| request_->set_request_context(request_context_getter_); |
| request_->set_upload_data("text/plain", list_data); |
| request_->Start(); |
| |
| // Begin the update request timeout. |
| update_timer_.Start(TimeDelta::FromSeconds(kSbMaxUpdateWaitSec), this, |
| &SafeBrowsingProtocolManager::UpdateResponseTimeout); |
| } |
| |
| // If we haven't heard back from the server with an update response, this method |
| // will run. Close the current update session and schedule another update. |
| void SafeBrowsingProtocolManager::UpdateResponseTimeout() { |
| DCHECK_EQ(request_type_, UPDATE_REQUEST); |
| request_.reset(); |
| UpdateFinished(false); |
| ScheduleNextUpdate(false); |
| } |
| |
| void SafeBrowsingProtocolManager::OnChunkInserted() { |
| chunk_pending_to_write_ = false; |
| |
| if (chunk_request_urls_.empty()) { |
| UMA_HISTOGRAM_LONG_TIMES("SB2.Update", Time::Now() - last_update_); |
| UpdateFinished(true); |
| } else { |
| IssueChunkRequest(); |
| } |
| } |
| |
| // Sends a SafeBrowsing "hit" for UMA users. |
| void SafeBrowsingProtocolManager::ReportSafeBrowsingHit( |
| const GURL& malicious_url, |
| const GURL& page_url, |
| const GURL& referrer_url, |
| bool is_subresource, |
| SafeBrowsingService::UrlCheckResult threat_type) { |
| GURL report_url = SafeBrowsingHitUrl(malicious_url, page_url, |
| referrer_url, is_subresource, |
| threat_type); |
| URLFetcher* report = new URLFetcher(report_url, URLFetcher::GET, this); |
| report->set_load_flags(net::LOAD_DISABLE_CACHE); |
| report->set_request_context(request_context_getter_); |
| report->Start(); |
| safebrowsing_reports_.insert(report); |
| } |
| |
| // Sends malware details for users who opt-in. |
| void SafeBrowsingProtocolManager::ReportMalwareDetails( |
| const std::string& report) { |
| GURL report_url = MalwareDetailsUrl(); |
| URLFetcher* fetcher = new URLFetcher(report_url, URLFetcher::POST, this); |
| fetcher->set_load_flags(net::LOAD_DISABLE_CACHE); |
| fetcher->set_request_context(request_context_getter_); |
| fetcher->set_upload_data("application/octet-stream", report); |
| // Don't try too hard to send reports on failures. |
| fetcher->set_automatically_retry_on_5xx(false); |
| fetcher->Start(); |
| safebrowsing_reports_.insert(fetcher); |
| } |
| |
| |
| // static |
| std::string SafeBrowsingProtocolManager::FormatList( |
| const SBListChunkRanges& list, bool use_mac) { |
| std::string formatted_results; |
| formatted_results.append(list.name); |
| formatted_results.append(";"); |
| if (!list.adds.empty()) { |
| formatted_results.append("a:" + list.adds); |
| if (!list.subs.empty() || use_mac) |
| formatted_results.append(":"); |
| } |
| if (!list.subs.empty()) { |
| formatted_results.append("s:" + list.subs); |
| if (use_mac) |
| formatted_results.append(":"); |
| } |
| if (use_mac) |
| formatted_results.append("mac"); |
| formatted_results.append("\n"); |
| |
| return formatted_results; |
| } |
| |
| void SafeBrowsingProtocolManager::HandleReKey() { |
| client_key_.clear(); |
| wrapped_key_.clear(); |
| IssueKeyRequest(); |
| } |
| |
| void SafeBrowsingProtocolManager::HandleGetHashError(const Time& now) { |
| int next = GetNextBackOffTime(&gethash_error_count_, &gethash_back_off_mult_); |
| next_gethash_time_ = now + TimeDelta::FromSeconds(next); |
| } |
| |
| void SafeBrowsingProtocolManager::UpdateFinished(bool success) { |
| UMA_HISTOGRAM_COUNTS("SB2.UpdateSize", update_size_); |
| update_size_ = 0; |
| sb_service_->UpdateFinished(success); |
| } |
| |
| std::string SafeBrowsingProtocolManager::ComposeUrl( |
| const std::string& prefix, const std::string& method, |
| const std::string& client_name, const std::string& version, |
| const std::string& additional_query) { |
| DCHECK(!prefix.empty() && !method.empty() && |
| !client_name.empty() && !version.empty()); |
| std::string url = StringPrintf("%s/%s?client=%s&appver=%s&pver=2.2", |
| prefix.c_str(), method.c_str(), |
| client_name.c_str(), version.c_str()); |
| if (!additional_query.empty()) { |
| DCHECK(url.find("?") != std::string::npos); |
| url.append("&"); |
| url.append(additional_query); |
| } |
| return url; |
| } |
| |
| GURL SafeBrowsingProtocolManager::UpdateUrl(bool use_mac) const { |
| std::string url = ComposeUrl(http_url_prefix_, "downloads", client_name_, |
| version_, additional_query_); |
| if (use_mac) { |
| url.append("&wrkey="); |
| url.append(wrapped_key_); |
| } |
| return GURL(url); |
| } |
| |
| GURL SafeBrowsingProtocolManager::GetHashUrl(bool use_mac) const { |
| std::string url= ComposeUrl(http_url_prefix_, "gethash", client_name_, |
| version_, additional_query_); |
| if (use_mac) { |
| url.append("&wrkey="); |
| url.append(wrapped_key_); |
| } |
| return GURL(url); |
| } |
| |
| GURL SafeBrowsingProtocolManager::MacKeyUrl() const { |
| return GURL(ComposeUrl(https_url_prefix_, "newkey", client_name_, version_, |
| additional_query_)); |
| } |
| |
| GURL SafeBrowsingProtocolManager::SafeBrowsingHitUrl( |
| const GURL& malicious_url, const GURL& page_url, |
| const GURL& referrer_url, bool is_subresource, |
| SafeBrowsingService::UrlCheckResult threat_type) const { |
| DCHECK(threat_type == SafeBrowsingService::URL_MALWARE || |
| threat_type == SafeBrowsingService::URL_PHISHING); |
| // The malware and phishing hits go over HTTP. |
| std::string url = ComposeUrl(http_url_prefix_, "report", client_name_, |
| version_, additional_query_); |
| return GURL(StringPrintf("%s&evts=%s&evtd=%s&evtr=%s&evhr=%s&evtb=%d", |
| url.c_str(), |
| (threat_type == SafeBrowsingService::URL_MALWARE) ? |
| "malblhit" : "phishblhit", |
| EscapeQueryParamValue(malicious_url.spec(), true).c_str(), |
| EscapeQueryParamValue(page_url.spec(), true).c_str(), |
| EscapeQueryParamValue(referrer_url.spec(), true).c_str(), |
| is_subresource)); |
| } |
| |
| GURL SafeBrowsingProtocolManager::MalwareDetailsUrl() const { |
| // The malware details go over HTTPS. |
| std::string url = StringPrintf( |
| "%s/clientreport/malware?client=%s&appver=%s&pver=1.0", |
| https_url_prefix_.c_str(), |
| client_name_.c_str(), |
| version_.c_str()); |
| return GURL(url); |
| } |
| |
| GURL SafeBrowsingProtocolManager::NextChunkUrl(const std::string& url) const { |
| std::string next_url; |
| if (!StartsWithASCII(url, "http://", false) && |
| !StartsWithASCII(url, "https://", false)) { |
| next_url.append("http://"); |
| next_url.append(url); |
| } else { |
| next_url = url; |
| } |
| if (!additional_query_.empty()) { |
| if (next_url.find("?") != std::string::npos) { |
| next_url.append("&"); |
| } else { |
| next_url.append("?"); |
| } |
| next_url.append(additional_query_); |
| } |
| return GURL(next_url); |
| } |