summaryrefslogtreecommitdiffstats
path: root/features/step_definitions/erase_memory.rb
diff options
context:
space:
mode:
authorHolger Levsen <holger@layer-acht.org>2014-12-19 02:03:30 +0100
committerHolger Levsen <holger@layer-acht.org>2014-12-21 09:47:01 +0100
commitbc9261df47a5dc59b670142485329299c944f984 (patch)
tree65cf0370e2438c14533bb7010ee66fc99cf91541 /features/step_definitions/erase_memory.rb
parent1bd026fa13481d9dcf14ab92baa4277324244466 (diff)
downloadjenkins.debian.net-bc9261df47a5dc59b670142485329299c944f984.tar.xz
lvc: remove some test scenarios we will never test (but are nice features in tails)
Diffstat (limited to 'features/step_definitions/erase_memory.rb')
-rw-r--r--features/step_definitions/erase_memory.rb172
1 files changed, 0 insertions, 172 deletions
diff --git a/features/step_definitions/erase_memory.rb b/features/step_definitions/erase_memory.rb
deleted file mode 100644
index 171f997c..00000000
--- a/features/step_definitions/erase_memory.rb
+++ /dev/null
@@ -1,172 +0,0 @@
-Given /^the computer is a modern 64-bit system$/ do
- next if @skip_steps_while_restoring_background
- @vm.set_arch("x86_64")
- @vm.drop_hypervisor_feature("nonpae")
- @vm.add_hypervisor_feature("pae")
-end
-
-Given /^the computer is an old pentium without the PAE extension$/ do
- next if @skip_steps_while_restoring_background
- @vm.set_arch("i686")
- @vm.drop_hypervisor_feature("pae")
- # libvirt claim the following feature doesn't exit even though
- # it's listed in the hvm i686 capabilities...
-# @vm.add_hypervisor_feature("nonpae")
- # ... so we use a workaround until we can figure this one out.
- @vm.disable_pae_workaround
-end
-
-def which_kernel
- kernel_path = @vm.execute("/usr/local/bin/tails-get-bootinfo kernel").stdout.chomp
- return File.basename(kernel_path)
-end
-
-Given /^the PAE kernel is running$/ do
- next if @skip_steps_while_restoring_background
- kernel = which_kernel
- assert_equal("vmlinuz2", kernel)
-end
-
-Given /^the non-PAE kernel is running$/ do
- next if @skip_steps_while_restoring_background
- kernel = which_kernel
- assert_equal("vmlinuz", kernel)
-end
-
-def used_ram_in_MiB
- return @vm.execute("free -m | awk '/^-\\/\\+ buffers\\/cache:/ { print $3 }'").stdout.chomp.to_i
-end
-
-def detected_ram_in_MiB
- return @vm.execute("free -m | awk '/^Mem:/ { print $2 }'").stdout.chomp.to_i
-end
-
-Given /^at least (\d+) ([[:alpha:]]+) of RAM was detected$/ do |min_ram, unit|
- @detected_ram_m = detected_ram_in_MiB
- next if @skip_steps_while_restoring_background
- puts "Detected #{@detected_ram_m} MiB of RAM"
- min_ram_m = convert_to_MiB(min_ram.to_i, unit)
- # All RAM will not be reported by `free`, so we allow a 196 MB gap
- gap = convert_to_MiB(196, "MiB")
- assert(@detected_ram_m + gap >= min_ram_m, "Didn't detect enough RAM")
-end
-
-def pattern_coverage_in_guest_ram
- dump = "#{$tmp_dir}/memdump"
- # Workaround: when dumping the guest's memory via core_dump(), libvirt
- # will create files that only root can read. We therefore pre-create
- # them with more permissible permissions, which libvirt will preserve
- # (although it will change ownership) so that the user running the
- # script can grep the dump for the fillram pattern, and delete it.
- if File.exist?(dump)
- File.delete(dump)
- end
- FileUtils.touch(dump)
- FileUtils.chmod(0666, dump)
- @vm.domain.core_dump(dump)
- patterns = IO.popen("grep -c 'wipe_didnt_work' #{dump}").gets.to_i
- File.delete dump
- # Pattern is 16 bytes long
- patterns_b = patterns*16
- patterns_m = convert_to_MiB(patterns_b, 'b')
- coverage = patterns_b.to_f/convert_to_bytes(@detected_ram_m.to_f, 'MiB')
- puts "Pattern coverage: #{"%.3f" % (coverage*100)}% (#{patterns_m} MiB)"
- return coverage
-end
-
-Given /^I fill the guest's memory with a known pattern(| without verifying)$/ do |dont_verify|
- verify = dont_verify.empty?
- next if @skip_steps_while_restoring_background
-
- # Free some more memory by dropping the caches etc.
- @vm.execute("echo 3 > /proc/sys/vm/drop_caches")
-
- # The (guest) kernel may freeze when approaching full memory without
- # adjusting the OOM killer and memory overcommitment limitations.
- [
- "echo 256 > /proc/sys/vm/min_free_kbytes",
- "echo 2 > /proc/sys/vm/overcommit_memory",
- "echo 97 > /proc/sys/vm/overcommit_ratio",
- "echo 1 > /proc/sys/vm/oom_kill_allocating_task",
- "echo 0 > /proc/sys/vm/oom_dump_tasks"
- ].each { |c| @vm.execute(c) }
-
- # The remote shell is sometimes OOM killed when we fill the memory,
- # and since we depend on it after the memory fill we try to prevent
- # that from happening.
- pid = @vm.pidof("autotest_remote_shell.py")[0]
- @vm.execute("echo -17 > /proc/#{pid}/oom_adj")
-
- used_mem_before_fill = used_ram_in_MiB
-
- # To be sure that we fill all memory we run one fillram instance
- # for each GiB of detected memory, rounded up. We also kill all instances
- # after the first one has finished, i.e. when the memory is full,
- # since the others otherwise may continue re-filling the same memory
- # unnecessarily.
- instances = (@detected_ram_m.to_f/(2**10)).ceil
- instances.times { @vm.spawn('/usr/local/sbin/fillram; killall fillram') }
- # We make sure that the filling has started...
- try_for(10, { :msg => "fillram didn't start" }) {
- @vm.has_process?("fillram")
- }
- STDERR.print "Memory fill progress: "
- ram_usage = ""
- remove_chars = 0
- # ... and that it finishes
- try_for(instances*2*60, { :msg => "fillram didn't complete, probably the VM crashed" }) do
- used_ram = used_ram_in_MiB
- remove_chars = ram_usage.size
- ram_usage = "%3d%% " % ((used_ram.to_f/@detected_ram_m)*100)
- STDERR.print "\b"*remove_chars + ram_usage
- ! @vm.has_process?("fillram")
- end
- STDERR.print "\b"*remove_chars + "finished.\n"
- if verify
- coverage = pattern_coverage_in_guest_ram()
- # Let's aim for having the pattern cover at least 80% of the free RAM.
- # More would be good, but it seems like OOM kill strikes around 90%,
- # and we don't want this test to fail all the time.
- min_coverage = ((@detected_ram_m - used_mem_before_fill).to_f /
- @detected_ram_m.to_f)*0.75
- assert(coverage > min_coverage,
- "#{"%.3f" % (coverage*100)}% of the memory is filled with the " +
- "pattern, but more than #{"%.3f" % (min_coverage*100)}% was expected")
- end
-end
-
-Then /^I find very few patterns in the guest's memory$/ do
- next if @skip_steps_while_restoring_background
- coverage = pattern_coverage_in_guest_ram()
- max_coverage = 0.005
- assert(coverage < max_coverage,
- "#{"%.3f" % (coverage*100)}% of the memory is filled with the " +
- "pattern, but less than #{"%.3f" % (max_coverage*100)}% was expected")
-end
-
-Then /^I find many patterns in the guest's memory$/ do
- next if @skip_steps_while_restoring_background
- coverage = pattern_coverage_in_guest_ram()
- min_coverage = 0.7
- assert(coverage > min_coverage,
- "#{"%.3f" % (coverage*100)}% of the memory is filled with the " +
- "pattern, but more than #{"%.3f" % (min_coverage*100)}% was expected")
-end
-
-When /^I reboot without wiping the memory$/ do
- next if @skip_steps_while_restoring_background
- @vm.reset
- @screen.wait('TailsBootSplashPostReset.png', 30)
-end
-
-When /^I shutdown and wait for Tails to finish wiping the memory$/ do
- next if @skip_steps_while_restoring_background
- @vm.execute("halt")
- nr_gibs_of_ram = (@detected_ram_m.to_f/(2**10)).ceil
- try_for(nr_gibs_of_ram*5*60, { :msg => "memory wipe didn't finish, probably the VM crashed" }) do
- # We spam keypresses to prevent console blanking from hiding the
- # image we're waiting for
- @screen.type(" ")
- @screen.wait('MemoryWipeCompleted.png')
- end
-end