Commit 04dd61be by Adam Palay

remove logging

parent a7d638de
...@@ -62,7 +62,6 @@ from importlib import import_module ...@@ -62,7 +62,6 @@ from importlib import import_module
from mongodb_proxy import autoretry_read from mongodb_proxy import autoretry_read
from path import Path as path from path import Path as path
from pytz import UTC from pytz import UTC
import traceback
from bson.objectid import ObjectId from bson.objectid import ObjectId
from xblock.core import XBlock from xblock.core import XBlock
...@@ -688,10 +687,6 @@ class SplitMongoModuleStore(SplitBulkWriteMixin, ModuleStoreWriteBase): ...@@ -688,10 +687,6 @@ class SplitMongoModuleStore(SplitBulkWriteMixin, ModuleStoreWriteBase):
if self.request_cache is not None: if self.request_cache is not None:
self.services["request_cache"] = self.request_cache self.services["request_cache"] = self.request_cache
# set the maximum number of courses we expect
# to see in the request cache
self.max_num_courses_in_cache = 10
self.signal_handler = signal_handler self.signal_handler = signal_handler
def close_connections(self): def close_connections(self):
...@@ -800,26 +795,6 @@ class SplitMongoModuleStore(SplitBulkWriteMixin, ModuleStoreWriteBase): ...@@ -800,26 +795,6 @@ class SplitMongoModuleStore(SplitBulkWriteMixin, ModuleStoreWriteBase):
""" """
if self.request_cache is not None: if self.request_cache is not None:
self.request_cache.data.setdefault('course_cache', {})[course_version_guid] = system self.request_cache.data.setdefault('course_cache', {})[course_version_guid] = system
number_courses_in_cache = len(self.request_cache.data['course_cache'])
if number_courses_in_cache > self.max_num_courses_in_cache:
# We shouldn't have any scenarios where there are many
# courses in the request cache. If there are, it's probably
# indicative of a leak. In this case, we should log that here
# with a stacktrace.
course_structure_ids = self.request_cache.data['course_cache'].keys()[:10]
log.warning(
(
"There are more than %d (%d) "
"courses in the request cache. This may be indicative "
"of a memory leak.\n\nHere are some of the course "
"structures' mongo ids that are now in the cache: %s\n\n"
"And here's the current stack:\n%s"
),
self.max_num_courses_in_cache,
number_courses_in_cache,
", ".join(unicode(structure_id) for structure_id in course_structure_ids),
"".join(traceback.format_stack()[-10:])
)
return system return system
def _clear_cache(self, course_version_guid=None): def _clear_cache(self, course_version_guid=None):
......
""" """
Test split modulestore w/o using any django stuff. Test split modulestore w/o using any django stuff.
""" """
from mock import patch, Mock from mock import patch
import datetime import datetime
from importlib import import_module from importlib import import_module
from path import Path as path from path import Path as path
...@@ -627,22 +627,6 @@ class SplitModuleCourseTests(SplitModuleTest): ...@@ -627,22 +627,6 @@ class SplitModuleCourseTests(SplitModuleTest):
self.assertEqual(course.edited_by, "testassist@edx.org") self.assertEqual(course.edited_by, "testassist@edx.org")
self.assertDictEqual(course.grade_cutoffs, {"Pass": 0.45}) self.assertDictEqual(course.grade_cutoffs, {"Pass": 0.45})
@ddt.data((10, 0), (2, 1))
@ddt.unpack
@patch("xmodule.modulestore.split_mongo.split.log.warning")
def test_request_cache_max_courses(self, max_courses, expected_warnings, mock_log):
"""
Test that we warn if there are too many courses in the request cache
at once.
"""
mock_request_cache = Mock()
mock_request_cache.data = {}
modulestore().max_num_courses_in_cache = max_courses
modulestore().request_cache = mock_request_cache
modulestore().get_courses(branch=BRANCH_NAME_DRAFT)
self.assertEqual(mock_log.call_count, expected_warnings)
@patch('xmodule.tabs.CourseTab.from_json', side_effect=mock_tab_from_json) @patch('xmodule.tabs.CourseTab.from_json', side_effect=mock_tab_from_json)
def test_get_courses_with_same_course_index(self, _from_json): def test_get_courses_with_same_course_index(self, _from_json):
""" """
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment