/*
- * Copyright (c) 2012, 2014 ARM Limited
+ * Copyright (c) 2012, 2014, 2018 ARM Limited
* All rights reserved
*
* The license below extends only to copyright in the software and shall
* Authors: Andreas Hansson
*/
+#include "mem/physical.hh"
+
+#include <fcntl.h>
#include <sys/mman.h>
#include <sys/types.h>
#include <sys/user.h>
-#include <fcntl.h>
#include <unistd.h>
#include <zlib.h>
#include "debug/AddrRanges.hh"
#include "debug/Checkpoint.hh"
#include "mem/abstract_mem.hh"
-#include "mem/physical.hh"
/**
* On Linux, MAP_NORESERVE allow us to simulate a very large memory
* without committing to actually providing the swap space on the
- * host. On OSX the MAP_NORESERVE flag does not exist, so simply make
- * it 0.
+ * host. On FreeBSD or OSX the MAP_NORESERVE flag does not exist,
+ * so simply make it 0.
*/
-#if defined(__APPLE__)
+#if defined(__APPLE__) || defined(__FreeBSD__)
#ifndef MAP_NORESERVE
#define MAP_NORESERVE 0
#endif
PhysicalMemory::PhysicalMemory(const string& _name,
const vector<AbstractMemory*>& _memories,
bool mmap_using_noreserve) :
- _name(_name), rangeCache(addrMap.end()), size(0),
- mmapUsingNoReserve(mmap_using_noreserve)
+ _name(_name), size(0), mmapUsingNoReserve(mmap_using_noreserve)
{
if (mmap_using_noreserve)
warn("Not reserving swap space. May cause SIGSEGV on actual usage\n");
// memories are allowed to overlap in the logic address
// map
vector<AbstractMemory*> unmapped_mems{m};
- createBackingStore(m->getAddrRange(), unmapped_mems);
+ createBackingStore(m->getAddrRange(), unmapped_mems,
+ m->isConfReported(), m->isInAddrMap(),
+ m->isKvmMap());
}
}
if (!intlv_ranges.empty() &&
!intlv_ranges.back().mergesWith(r.first)) {
AddrRange merged_range(intlv_ranges);
- createBackingStore(merged_range, curr_memories);
+
+ AbstractMemory *f = curr_memories.front();
+ for (const auto& c : curr_memories)
+ if (f->isConfReported() != c->isConfReported() ||
+ f->isInAddrMap() != c->isInAddrMap() ||
+ f->isKvmMap() != c->isKvmMap())
+ fatal("Inconsistent flags in an interleaved "
+ "range\n");
+
+ createBackingStore(merged_range, curr_memories,
+ f->isConfReported(), f->isInAddrMap(),
+ f->isKvmMap());
+
intlv_ranges.clear();
curr_memories.clear();
}
curr_memories.push_back(r.second);
} else {
vector<AbstractMemory*> single_memory{r.second};
- createBackingStore(r.first, single_memory);
+ createBackingStore(r.first, single_memory,
+ r.second->isConfReported(),
+ r.second->isInAddrMap(),
+ r.second->isKvmMap());
}
}
}
// ahead and do it
if (!intlv_ranges.empty()) {
AddrRange merged_range(intlv_ranges);
- createBackingStore(merged_range, curr_memories);
+
+ AbstractMemory *f = curr_memories.front();
+ for (const auto& c : curr_memories)
+ if (f->isConfReported() != c->isConfReported() ||
+ f->isInAddrMap() != c->isInAddrMap() ||
+ f->isKvmMap() != c->isKvmMap())
+ fatal("Inconsistent flags in an interleaved "
+ "range\n");
+
+ createBackingStore(merged_range, curr_memories,
+ f->isConfReported(), f->isInAddrMap(),
+ f->isKvmMap());
}
}
void
PhysicalMemory::createBackingStore(AddrRange range,
- const vector<AbstractMemory*>& _memories)
+ const vector<AbstractMemory*>& _memories,
+ bool conf_table_reported,
+ bool in_addr_map, bool kvm_map)
{
panic_if(range.interleaved(),
"Cannot create backing store for interleaved range %s\n",
// remember this backing store so we can checkpoint it and unmap
// it appropriately
- backingStore.push_back(make_pair(range, pmem));
+ backingStore.emplace_back(range, pmem,
+ conf_table_reported, in_addr_map, kvm_map);
// point the memories to their backing store
for (const auto& m : _memories) {
{
// unmap the backing store
for (auto& s : backingStore)
- munmap((char*)s.second, s.first.size());
+ munmap((char*)s.pmem, s.range.size());
}
bool
PhysicalMemory::isMemAddr(Addr addr) const
{
- // see if the address is within the last matched range
- if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
- return true;
- } else {
- // lookup in the interval tree
- const auto& r = addrMap.find(addr);
- if (r == addrMap.end()) {
- // not in the cache, and not in the tree
- return false;
- }
- // the range is in the tree, update the cache
- rangeCache = r;
- return true;
- }
+ return addrMap.contains(addr) != addrMap.end();
}
AddrRangeList
PhysicalMemory::access(PacketPtr pkt)
{
assert(pkt->isRequest());
- Addr addr = pkt->getAddr();
- if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
- rangeCache->second->access(pkt);
- } else {
- // do not update the cache here, as we typically call
- // isMemAddr before calling access
- const auto& m = addrMap.find(addr);
- assert(m != addrMap.end());
- m->second->access(pkt);
- }
+ const auto& m = addrMap.contains(pkt->getAddrRange());
+ assert(m != addrMap.end());
+ m->second->access(pkt);
}
void
PhysicalMemory::functionalAccess(PacketPtr pkt)
{
assert(pkt->isRequest());
- Addr addr = pkt->getAddr();
- if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
- rangeCache->second->functionalAccess(pkt);
- } else {
- // do not update the cache here, as we typically call
- // isMemAddr before calling functionalAccess
- const auto& m = addrMap.find(addr);
- assert(m != addrMap.end());
- m->second->functionalAccess(pkt);
- }
+ const auto& m = addrMap.contains(pkt->getAddrRange());
+ assert(m != addrMap.end());
+ m->second->functionalAccess(pkt);
}
void
// store each backing store memory segment in a file
for (auto& s : backingStore) {
ScopedCheckpointSection sec(cp, csprintf("store%d", store_id));
- serializeStore(cp, store_id++, s.first, s.second);
+ serializeStore(cp, store_id++, s.range, s.pmem);
}
}
vector<ContextID> lal_cid;
UNSERIALIZE_CONTAINER(lal_addr);
UNSERIALIZE_CONTAINER(lal_cid);
- for(size_t i = 0; i < lal_addr.size(); ++i) {
- const auto& m = addrMap.find(lal_addr[i]);
+ for (size_t i = 0; i < lal_addr.size(); ++i) {
+ const auto& m = addrMap.contains(lal_addr[i]);
m->second->addLockedAddr(LockedAddr(lal_addr[i], lal_cid[i]));
}
fatal("Can't open physical memory checkpoint file '%s'", filename);
// we've already got the actual backing store mapped
- uint8_t* pmem = backingStore[store_id].second;
- AddrRange range = backingStore[store_id].first;
+ uint8_t* pmem = backingStore[store_id].pmem;
+ AddrRange range = backingStore[store_id].range;
long range_size;
UNSERIALIZE_SCALAR(range_size);