
For convenience, we had added the folder that dotest.py was in to sys.path, so that we could easily write things like `import lldbutil` from anywhere and any test. This introduces a subtle problem when using Python's package system, because when unittest2 imports a particular test suite, the test suite is detached from the package. Thus, writing "import lldbutil" from dotest imports it as part of the package, and writing the same line from a test does a fresh import since the importing module was not part of the same package. The real way to fix this is to use absolute imports everywhere. Instead of writing "import lldbutil", we need to write "import lldbsuite.test.util". This patch fixes up that and all other similar cases, and additionally removes the script directory from sys.path to ensure that this can't happen again. llvm-svn: 251886
202 lines
7.2 KiB
Python
202 lines
7.2 KiB
Python
from __future__ import print_function
|
|
|
|
import use_lldb_suite
|
|
|
|
import gdbremote_testcase
|
|
from lldbsuite.test.lldbtest import *
|
|
|
|
class TestGdbRemoteAuxvSupport(gdbremote_testcase.GdbRemoteTestCaseBase):
|
|
|
|
mydir = TestBase.compute_mydir(__file__)
|
|
|
|
AUXV_SUPPORT_FEATURE_NAME = "qXfer:auxv:read"
|
|
|
|
def has_auxv_support(self):
|
|
inferior_args = ["message:main entered", "sleep:5"]
|
|
procs = self.prep_debug_monitor_and_inferior(inferior_args=inferior_args)
|
|
|
|
# Don't do anything until we match the launched inferior main entry output.
|
|
# Then immediately interrupt the process.
|
|
# This prevents auxv data being asked for before it's ready and leaves
|
|
# us in a stopped state.
|
|
self.test_sequence.add_log_lines([
|
|
# Start the inferior...
|
|
"read packet: $c#63",
|
|
# ... match output....
|
|
{ "type":"output_match", "regex":r"^message:main entered\r\n$" },
|
|
], True)
|
|
# ... then interrupt.
|
|
self.add_interrupt_packets()
|
|
self.add_qSupported_packets()
|
|
|
|
context = self.expect_gdbremote_sequence()
|
|
self.assertIsNotNone(context)
|
|
|
|
features = self.parse_qSupported_response(context)
|
|
return self.AUXV_SUPPORT_FEATURE_NAME in features and features[self.AUXV_SUPPORT_FEATURE_NAME] == "+"
|
|
|
|
def get_raw_auxv_data(self):
|
|
# Start up llgs and inferior, and check for auxv support.
|
|
if not self.has_auxv_support():
|
|
self.skipTest("auxv data not supported")
|
|
|
|
# Grab pointer size for target. We'll assume that is equivalent to an unsigned long on the target.
|
|
# Auxv is specified in terms of pairs of unsigned longs.
|
|
self.reset_test_sequence()
|
|
self.add_process_info_collection_packets()
|
|
|
|
context = self.expect_gdbremote_sequence()
|
|
self.assertIsNotNone(context)
|
|
|
|
proc_info = self.parse_process_info_response(context)
|
|
self.assertIsNotNone(proc_info)
|
|
self.assertTrue("ptrsize" in proc_info)
|
|
word_size = int(proc_info["ptrsize"])
|
|
|
|
OFFSET = 0
|
|
LENGTH = 0x400
|
|
|
|
# Grab the auxv data.
|
|
self.reset_test_sequence()
|
|
self.test_sequence.add_log_lines([
|
|
"read packet: $qXfer:auxv:read::{:x},{:x}:#00".format(OFFSET, LENGTH),
|
|
{"direction":"send", "regex":re.compile(r"^\$([^E])(.*)#[0-9a-fA-F]{2}$", re.MULTILINE|re.DOTALL), "capture":{1:"response_type", 2:"content_raw"} }
|
|
], True)
|
|
|
|
context = self.expect_gdbremote_sequence()
|
|
self.assertIsNotNone(context)
|
|
|
|
# Ensure we end up with all auxv data in one packet.
|
|
# FIXME don't assume it all comes back in one packet.
|
|
self.assertEqual(context.get("response_type"), "l")
|
|
|
|
# Decode binary data.
|
|
content_raw = context.get("content_raw")
|
|
self.assertIsNotNone(content_raw)
|
|
return (word_size, self.decode_gdbremote_binary(content_raw))
|
|
|
|
def supports_auxv(self):
|
|
# When non-auxv platforms support llgs, skip the test on platforms
|
|
# that don't support auxv.
|
|
self.assertTrue(self.has_auxv_support())
|
|
|
|
#
|
|
# We skip the "supports_auxv" test on debugserver. The rest of the tests
|
|
# appropriately skip the auxv tests if the support flag is not present
|
|
# in the qSupported response, so the debugserver test bits are still there
|
|
# in case debugserver code one day does have auxv support and thus those
|
|
# tests don't get skipped.
|
|
#
|
|
|
|
@llgs_test
|
|
def test_supports_auxv_llgs(self):
|
|
self.init_llgs_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.supports_auxv()
|
|
|
|
def auxv_data_is_correct_size(self):
|
|
(word_size, auxv_data) = self.get_raw_auxv_data()
|
|
self.assertIsNotNone(auxv_data)
|
|
|
|
# Ensure auxv data is a multiple of 2*word_size (there should be two unsigned long fields per auxv entry).
|
|
self.assertEqual(len(auxv_data) % (2*word_size), 0)
|
|
# print("auxv contains {} entries".format(len(auxv_data) / (2*word_size)))
|
|
|
|
@debugserver_test
|
|
def test_auxv_data_is_correct_size_debugserver(self):
|
|
self.init_debugserver_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_data_is_correct_size()
|
|
|
|
@llgs_test
|
|
def test_auxv_data_is_correct_size_llgs(self):
|
|
self.init_llgs_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_data_is_correct_size()
|
|
|
|
def auxv_keys_look_valid(self):
|
|
(word_size, auxv_data) = self.get_raw_auxv_data()
|
|
self.assertIsNotNone(auxv_data)
|
|
|
|
# Grab endian.
|
|
self.reset_test_sequence()
|
|
self.add_process_info_collection_packets()
|
|
context = self.expect_gdbremote_sequence()
|
|
self.assertIsNotNone(context)
|
|
|
|
process_info = self.parse_process_info_response(context)
|
|
self.assertIsNotNone(process_info)
|
|
endian = process_info.get("endian")
|
|
self.assertIsNotNone(endian)
|
|
|
|
auxv_dict = self.build_auxv_dict(endian, word_size, auxv_data)
|
|
self.assertIsNotNone(auxv_dict)
|
|
|
|
# Verify keys look reasonable.
|
|
for auxv_key in auxv_dict:
|
|
self.assertTrue(auxv_key >= 1)
|
|
self.assertTrue(auxv_key <= 1000)
|
|
# print("auxv dict: {}".format(auxv_dict))
|
|
|
|
@debugserver_test
|
|
def test_auxv_keys_look_valid_debugserver(self):
|
|
self.init_debugserver_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_keys_look_valid()
|
|
|
|
@llgs_test
|
|
def test_auxv_keys_look_valid_llgs(self):
|
|
self.init_llgs_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_keys_look_valid()
|
|
|
|
def auxv_chunked_reads_work(self):
|
|
# Verify that multiple smaller offset,length reads of auxv data
|
|
# return the same data as a single larger read.
|
|
|
|
# Grab the auxv data with a single large read here.
|
|
(word_size, auxv_data) = self.get_raw_auxv_data()
|
|
self.assertIsNotNone(auxv_data)
|
|
|
|
# Grab endian.
|
|
self.reset_test_sequence()
|
|
self.add_process_info_collection_packets()
|
|
context = self.expect_gdbremote_sequence()
|
|
self.assertIsNotNone(context)
|
|
|
|
process_info = self.parse_process_info_response(context)
|
|
self.assertIsNotNone(process_info)
|
|
endian = process_info.get("endian")
|
|
self.assertIsNotNone(endian)
|
|
|
|
auxv_dict = self.build_auxv_dict(endian, word_size, auxv_data)
|
|
self.assertIsNotNone(auxv_dict)
|
|
|
|
iterated_auxv_data = self.read_binary_data_in_chunks("qXfer:auxv:read::", 2*word_size)
|
|
self.assertIsNotNone(iterated_auxv_data)
|
|
|
|
auxv_dict_iterated = self.build_auxv_dict(endian, word_size, iterated_auxv_data)
|
|
self.assertIsNotNone(auxv_dict_iterated)
|
|
|
|
# Verify both types of data collection returned same content.
|
|
self.assertEqual(auxv_dict_iterated, auxv_dict)
|
|
|
|
@debugserver_test
|
|
def test_auxv_chunked_reads_work_debugserver(self):
|
|
self.init_debugserver_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_chunked_reads_work()
|
|
|
|
@llgs_test
|
|
def test_auxv_chunked_reads_work_llgs(self):
|
|
self.init_llgs_test()
|
|
self.build()
|
|
self.set_inferior_startup_launch()
|
|
self.auxv_chunked_reads_work()
|