2 * Copyright (c) 2014 Advanced Micro Devices, Inc.
3 * Copyright (c) 2003 The Regents of The University of Michigan
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are
8 * met: redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer;
10 * redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution;
13 * neither the name of the copyright holders nor the names of its
14 * contributors may be used to endorse or promote products derived from
15 * this software without specific prior written permission.
17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29 * Authors: Steve Reinhardt
36 * Definitions of functional page table.
38 #include "mem/page_table.hh"
45 #include "base/bitfield.hh"
46 #include "base/intmath.hh"
47 #include "base/trace.hh"
48 #include "config/the_isa.hh"
49 #include "debug/MMU.hh"
50 #include "sim/faults.hh"
51 #include "sim/sim_object.hh"
54 using namespace TheISA
;
56 FuncPageTable::FuncPageTable(const std::string
&__name
,
57 uint64_t _pid
, Addr _pageSize
)
58 : PageTableBase(__name
, _pid
, _pageSize
)
62 FuncPageTable::~FuncPageTable()
67 FuncPageTable::map(Addr vaddr
, Addr paddr
, int64_t size
, uint64_t flags
)
69 bool clobber
= flags
& Clobber
;
70 // starting address must be page aligned
71 assert(pageOffset(vaddr
) == 0);
73 DPRINTF(MMU
, "Allocating Page: %#x-%#x\n", vaddr
, vaddr
+ size
);
75 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
, paddr
+= pageSize
) {
76 if (!clobber
&& (pTable
.find(vaddr
) != pTable
.end())) {
78 fatal("FuncPageTable::allocate: addr 0x%x already mapped", vaddr
);
81 pTable
[vaddr
] = TheISA::TlbEntry(pid
, vaddr
, paddr
,
84 eraseCacheEntry(vaddr
);
85 updateCache(vaddr
, pTable
[vaddr
]);
90 FuncPageTable::remap(Addr vaddr
, int64_t size
, Addr new_vaddr
)
92 assert(pageOffset(vaddr
) == 0);
93 assert(pageOffset(new_vaddr
) == 0);
95 DPRINTF(MMU
, "moving pages from vaddr %08p to %08p, size = %d\n", vaddr
,
99 size
-= pageSize
, vaddr
+= pageSize
, new_vaddr
+= pageSize
)
101 assert(pTable
.find(vaddr
) != pTable
.end());
103 pTable
[new_vaddr
] = pTable
[vaddr
];
105 eraseCacheEntry(vaddr
);
106 pTable
[new_vaddr
].updateVaddr(new_vaddr
);
107 updateCache(new_vaddr
, pTable
[new_vaddr
]);
112 FuncPageTable::unmap(Addr vaddr
, int64_t size
)
114 assert(pageOffset(vaddr
) == 0);
116 DPRINTF(MMU
, "Unmapping page: %#x-%#x\n", vaddr
, vaddr
+ size
);
118 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
) {
119 assert(pTable
.find(vaddr
) != pTable
.end());
121 eraseCacheEntry(vaddr
);
127 FuncPageTable::isUnmapped(Addr vaddr
, int64_t size
)
129 // starting address must be page aligned
130 assert(pageOffset(vaddr
) == 0);
132 for (; size
> 0; size
-= pageSize
, vaddr
+= pageSize
) {
133 if (pTable
.find(vaddr
) != pTable
.end()) {
142 FuncPageTable::lookup(Addr vaddr
, TheISA::TlbEntry
&entry
)
144 Addr page_addr
= pageAlign(vaddr
);
146 if (pTableCache
[0].valid
&& pTableCache
[0].vaddr
== page_addr
) {
147 entry
= pTableCache
[0].entry
;
150 if (pTableCache
[1].valid
&& pTableCache
[1].vaddr
== page_addr
) {
151 entry
= pTableCache
[1].entry
;
154 if (pTableCache
[2].valid
&& pTableCache
[2].vaddr
== page_addr
) {
155 entry
= pTableCache
[2].entry
;
159 PTableItr iter
= pTable
.find(page_addr
);
161 if (iter
== pTable
.end()) {
165 updateCache(page_addr
, iter
->second
);
166 entry
= iter
->second
;
171 PageTableBase::translate(Addr vaddr
, Addr
&paddr
)
173 TheISA::TlbEntry entry
;
174 if (!lookup(vaddr
, entry
)) {
175 DPRINTF(MMU
, "Couldn't Translate: %#x\n", vaddr
);
178 paddr
= pageOffset(vaddr
) + entry
.pageStart();
179 DPRINTF(MMU
, "Translating: %#x->%#x\n", vaddr
, paddr
);
184 PageTableBase::translate(RequestPtr req
)
187 assert(pageAlign(req
->getVaddr() + req
->getSize() - 1)
188 == pageAlign(req
->getVaddr()));
189 if (!translate(req
->getVaddr(), paddr
)) {
190 return Fault(new GenericPageTableFault(req
->getVaddr()));
192 req
->setPaddr(paddr
);
193 if ((paddr
& (pageSize
- 1)) + req
->getSize() > pageSize
) {
194 panic("Request spans page boundaries!\n");
201 FuncPageTable::serialize(CheckpointOut
&cp
) const
203 paramOut(cp
, "ptable.size", pTable
.size());
205 PTable::size_type count
= 0;
206 for (auto &pte
: pTable
) {
207 ScopedCheckpointSection
sec(cp
, csprintf("Entry%d", count
++));
209 paramOut(cp
, "vaddr", pte
.first
);
210 pte
.second
.serialize(cp
);
212 assert(count
== pTable
.size());
216 FuncPageTable::unserialize(CheckpointIn
&cp
)
219 paramIn(cp
, "ptable.size", count
);
221 for (int i
= 0; i
< count
; ++i
) {
222 ScopedCheckpointSection
sec(cp
, csprintf("Entry%d", i
));
224 std::unique_ptr
<TheISA::TlbEntry
> entry
;
227 paramIn(cp
, "vaddr", vaddr
);
228 entry
.reset(new TheISA::TlbEntry());
229 entry
->unserialize(cp
);
231 pTable
[vaddr
] = *entry
;