First customization to avm99963-bugs

- Only a service is used (default) in order to keep usage within Google
  Cloud's free tier.
- Settings have been adapted.
- Only a single DB instance is used (primary).
- Some cron jobs are executed less often.
- |threadsafe| has been set to "yes" so the single instance can serve
  concurrent requests (otherwise the search/list issues function
  breaks).

GitOrigin-RevId: 52130750a30a5aa2234a6dd3f5ae23c3fb1b8343
diff --git a/search/frontendsearchpipeline.py b/search/frontendsearchpipeline.py
index 367c52f..ec0a28e 100644
--- a/search/frontendsearchpipeline.py
+++ b/search/frontendsearchpipeline.py
@@ -951,8 +951,8 @@
     UserRPC for the created RPC call.
   """
   shard_id, subquery = shard_key
-  backend_host = modules.get_hostname(module='besearch')
-  url = 'http://%s%s' % (
+  backend_host = modules.get_hostname(module='default')
+  url = 'https://%s%s' % (
       backend_host,
       framework_helpers.FormatURL(
           [],
@@ -981,8 +981,8 @@
     project_id, logged_in_user_id, shard_id, invalidation_timestep,
     deadline=None, failfast=True):
   """Ask a backend to query one shard of the database."""
-  backend_host = modules.get_hostname(module='besearch')
-  url = 'http://%s%s' % (backend_host, framework_helpers.FormatURL(
+  backend_host = modules.get_hostname(module='default')
+  url = 'https://%s%s' % (backend_host, framework_helpers.FormatURL(
       None, urls.BACKEND_NONVIEWABLE,
       project_id=project_id or '',
       logged_in_user_id=logged_in_user_id or '',
diff --git a/search/test/frontendsearchpipeline_test.py b/search/test/frontendsearchpipeline_test.py
index b2e7fb3..432a9d1 100644
--- a/search/test/frontendsearchpipeline_test.py
+++ b/search/test/frontendsearchpipeline_test.py
@@ -937,7 +937,7 @@
     a_fake_rpc = testing_helpers.Blank(callback=None)
     urlfetch.create_rpc(deadline=settings.backend_deadline).AndReturn(
       a_fake_rpc)
-    modules.get_hostname(module='besearch')
+    modules.get_hostname(module='default')
     urlfetch.make_fetch_call(
       a_fake_rpc, mox.StrContains(
           urls.BACKEND_SEARCH + '?groupby=cc&invalidation_timestep=12345&'
@@ -967,7 +967,7 @@
     a_fake_rpc = testing_helpers.Blank(callback=None)
     urlfetch.create_rpc(deadline=settings.backend_deadline).AndReturn(
       a_fake_rpc)
-    modules.get_hostname(module='besearch')
+    modules.get_hostname(module='default')
     urlfetch.make_fetch_call(
         a_fake_rpc,
         mox.StrContains(
@@ -1002,7 +1002,7 @@
     a_fake_rpc = testing_helpers.Blank(callback=None)
     urlfetch.create_rpc(deadline=settings.backend_deadline).AndReturn(
       a_fake_rpc)
-    modules.get_hostname(module='besearch')
+    modules.get_hostname(module='default')
     urlfetch.make_fetch_call(
       a_fake_rpc, mox.StrContains(urls.BACKEND_NONVIEWABLE),
       follow_redirects=False, headers=mox.IsA(dict))