2 * Copyright (c) 2014 Advanced Micro Devices, Inc.
3 * Copyright (c) 2003 The Regents of The University of Michigan
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are
8 * met: redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer;
10 * redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution;
13 * neither the name of the copyright holders nor the names of its
14 * contributors may be used to endorse or promote products derived from
15 * this software without specific prior written permission.
17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29 * Authors: Steve Reinhardt
36 * Definitions of functional page table.
43 #include "base/bitfield.hh"
44 #include "base/intmath.hh"
45 #include "base/trace.hh"
46 #include "config/the_isa.hh"
47 #include "debug/MMU.hh"
48 #include "mem/page_table.hh"
49 #include "sim/faults.hh"
50 #include "sim/sim_object.hh"
53 using namespace TheISA
;
55 FuncPageTable::FuncPageTable(const std::string
&__name
,
56 uint64_t _pid
, Addr _pageSize
)
57 : PageTableBase(__name
, _pid
, _pageSize
)
61 FuncPageTable::~FuncPageTable()
66 FuncPageTable::map(Addr vaddr
, Addr paddr
, int64_t size
, uint64_t flags
)
68 bool clobber
= flags
& Clobber
;
69 // starting address must be page aligned
70 assert(pageOffset(vaddr
) == 0);
72 DPRINTF(MMU
, "Allocating Page: %#x-%#x\n", vaddr
, vaddr
+ size
);
74 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
, paddr
+= pageSize
) {
75 if (!clobber
&& (pTable
.find(vaddr
) != pTable
.end())) {
77 fatal("FuncPageTable::allocate: addr 0x%x already mapped", vaddr
);
80 pTable
[vaddr
] = TheISA::TlbEntry(pid
, vaddr
, paddr
,
83 eraseCacheEntry(vaddr
);
84 updateCache(vaddr
, pTable
[vaddr
]);
89 FuncPageTable::remap(Addr vaddr
, int64_t size
, Addr new_vaddr
)
91 assert(pageOffset(vaddr
) == 0);
92 assert(pageOffset(new_vaddr
) == 0);
94 DPRINTF(MMU
, "moving pages from vaddr %08p to %08p, size = %d\n", vaddr
,
98 size
-= pageSize
, vaddr
+= pageSize
, new_vaddr
+= pageSize
)
100 assert(pTable
.find(vaddr
) != pTable
.end());
102 pTable
[new_vaddr
] = pTable
[vaddr
];
104 eraseCacheEntry(vaddr
);
105 pTable
[new_vaddr
].updateVaddr(new_vaddr
);
106 updateCache(new_vaddr
, pTable
[new_vaddr
]);
111 FuncPageTable::unmap(Addr vaddr
, int64_t size
)
113 assert(pageOffset(vaddr
) == 0);
115 DPRINTF(MMU
, "Unmapping page: %#x-%#x\n", vaddr
, vaddr
+ size
);
117 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
) {
118 assert(pTable
.find(vaddr
) != pTable
.end());
120 eraseCacheEntry(vaddr
);
126 FuncPageTable::isUnmapped(Addr vaddr
, int64_t size
)
128 // starting address must be page aligned
129 assert(pageOffset(vaddr
) == 0);
131 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
) {
132 if (pTable
.find(vaddr
) != pTable
.end()) {
141 FuncPageTable::lookup(Addr vaddr
, TheISA::TlbEntry
&entry
)
143 Addr page_addr
= pageAlign(vaddr
);
145 if (pTableCache
[0].valid
&& pTableCache
[0].vaddr
== page_addr
) {
146 entry
= pTableCache
[0].entry
;
149 if (pTableCache
[1].valid
&& pTableCache
[1].vaddr
== page_addr
) {
150 entry
= pTableCache
[1].entry
;
153 if (pTableCache
[2].valid
&& pTableCache
[2].vaddr
== page_addr
) {
154 entry
= pTableCache
[2].entry
;
158 PTableItr iter
= pTable
.find(page_addr
);
160 if (iter
== pTable
.end()) {
164 updateCache(page_addr
, iter
->second
);
165 entry
= iter
->second
;
170 PageTableBase::translate(Addr vaddr
, Addr
&paddr
)
172 TheISA::TlbEntry entry
;
173 if (!lookup(vaddr
, entry
)) {
174 DPRINTF(MMU
, "Couldn't Translate: %#x\n", vaddr
);
177 paddr
= pageOffset(vaddr
) + entry
.pageStart();
178 DPRINTF(MMU
, "Translating: %#x->%#x\n", vaddr
, paddr
);
183 PageTableBase::translate(RequestPtr req
)
186 assert(pageAlign(req
->getVaddr() + req
->getSize() - 1)
187 == pageAlign(req
->getVaddr()));
188 if (!translate(req
->getVaddr(), paddr
)) {
189 return Fault(new GenericPageTableFault(req
->getVaddr()));
191 req
->setPaddr(paddr
);
192 if ((paddr
& (pageSize
- 1)) + req
->getSize() > pageSize
) {
193 panic("Request spans page boundaries!\n");
200 FuncPageTable::serialize(CheckpointOut
&cp
) const
202 paramOut(cp
, "ptable.size", pTable
.size());
204 PTable::size_type count
= 0;
205 for (auto &pte
: pTable
) {
206 ScopedCheckpointSection
sec(cp
, csprintf("Entry%d", count
++));
208 paramOut(cp
, "vaddr", pte
.first
);
209 pte
.second
.serialize(cp
);
211 assert(count
== pTable
.size());
215 FuncPageTable::unserialize(CheckpointIn
&cp
)
218 paramIn(cp
, "ptable.size", count
);
220 for (int i
= 0; i
< count
; ++i
) {
221 ScopedCheckpointSection
sec(cp
, csprintf("Entry%d", i
));
223 std::unique_ptr
<TheISA::TlbEntry
> entry
;
226 paramIn(cp
, "vaddr", vaddr
);
227 entry
.reset(new TheISA::TlbEntry());
228 entry
->unserialize(cp
);
230 pTable
[vaddr
] = *entry
;